var/home/core/zuul-output/0000755000175000017500000000000015113772752014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114007135015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005020241215114007126017666 0ustar rootrootDec 03 08:37:58 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 08:37:58 crc restorecon[4572]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:58 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 08:37:59 crc restorecon[4572]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 08:37:59 crc kubenswrapper[4573]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 08:37:59 crc kubenswrapper[4573]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 08:37:59 crc kubenswrapper[4573]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 08:37:59 crc kubenswrapper[4573]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 08:37:59 crc kubenswrapper[4573]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 08:37:59 crc kubenswrapper[4573]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.729244 4573 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733289 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733313 4573 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733318 4573 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733324 4573 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733329 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733333 4573 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733343 4573 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733347 4573 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733414 4573 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733421 4573 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733425 4573 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733436 4573 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733440 4573 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733444 4573 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733447 4573 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733452 4573 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733456 4573 feature_gate.go:330] unrecognized feature gate: Example Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733460 4573 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733464 4573 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733472 4573 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733476 4573 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733483 4573 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733489 4573 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733494 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733498 4573 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733503 4573 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733506 4573 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733510 4573 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733514 4573 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733518 4573 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733522 4573 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733529 4573 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733536 4573 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733542 4573 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733546 4573 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733550 4573 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733558 4573 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733564 4573 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733568 4573 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733572 4573 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733577 4573 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733639 4573 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733649 4573 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733655 4573 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733659 4573 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733663 4573 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733667 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733671 4573 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733675 4573 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733679 4573 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733684 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733688 4573 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733692 4573 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733697 4573 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733701 4573 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733709 4573 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733713 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733717 4573 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733724 4573 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733729 4573 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733733 4573 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733737 4573 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733742 4573 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733746 4573 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733750 4573 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733755 4573 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733763 4573 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733767 4573 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733770 4573 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733775 4573 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.733778 4573 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.733987 4573 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734002 4573 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734092 4573 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734101 4573 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734107 4573 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734112 4573 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734201 4573 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734210 4573 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734221 4573 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734227 4573 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734233 4573 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734237 4573 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734242 4573 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734248 4573 flags.go:64] FLAG: --cgroup-root="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734252 4573 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734260 4573 flags.go:64] FLAG: --client-ca-file="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734265 4573 flags.go:64] FLAG: --cloud-config="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734270 4573 flags.go:64] FLAG: --cloud-provider="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734276 4573 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734310 4573 flags.go:64] FLAG: --cluster-domain="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734315 4573 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734319 4573 flags.go:64] FLAG: --config-dir="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734323 4573 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734329 4573 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734339 4573 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734345 4573 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734350 4573 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734356 4573 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734360 4573 flags.go:64] FLAG: --contention-profiling="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734366 4573 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734375 4573 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734380 4573 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734385 4573 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734391 4573 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734397 4573 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734402 4573 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734407 4573 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734411 4573 flags.go:64] FLAG: --enable-server="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734416 4573 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734431 4573 flags.go:64] FLAG: --event-burst="100" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734436 4573 flags.go:64] FLAG: --event-qps="50" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734440 4573 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734452 4573 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734457 4573 flags.go:64] FLAG: --eviction-hard="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734463 4573 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734468 4573 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734473 4573 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734481 4573 flags.go:64] FLAG: --eviction-soft="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734485 4573 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734490 4573 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734495 4573 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734499 4573 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734504 4573 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734508 4573 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734513 4573 flags.go:64] FLAG: --feature-gates="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734521 4573 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734526 4573 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734531 4573 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734536 4573 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734540 4573 flags.go:64] FLAG: --healthz-port="10248" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734545 4573 flags.go:64] FLAG: --help="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734549 4573 flags.go:64] FLAG: --hostname-override="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734553 4573 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734558 4573 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734565 4573 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734570 4573 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734575 4573 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734580 4573 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734584 4573 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734588 4573 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734593 4573 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734597 4573 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734602 4573 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734609 4573 flags.go:64] FLAG: --kube-reserved="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734613 4573 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734617 4573 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734622 4573 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734626 4573 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734640 4573 flags.go:64] FLAG: --lock-file="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734648 4573 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734653 4573 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734660 4573 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734668 4573 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734672 4573 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734677 4573 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734682 4573 flags.go:64] FLAG: --logging-format="text" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734687 4573 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734692 4573 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734697 4573 flags.go:64] FLAG: --manifest-url="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734704 4573 flags.go:64] FLAG: --manifest-url-header="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734712 4573 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734717 4573 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734724 4573 flags.go:64] FLAG: --max-pods="110" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734729 4573 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734734 4573 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734739 4573 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734743 4573 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734749 4573 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734837 4573 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734842 4573 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734856 4573 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734862 4573 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734867 4573 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734873 4573 flags.go:64] FLAG: --pod-cidr="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734881 4573 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734904 4573 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734909 4573 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734915 4573 flags.go:64] FLAG: --pods-per-core="0" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734920 4573 flags.go:64] FLAG: --port="10250" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734924 4573 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734929 4573 flags.go:64] FLAG: --provider-id="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734935 4573 flags.go:64] FLAG: --qos-reserved="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734943 4573 flags.go:64] FLAG: --read-only-port="10255" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734948 4573 flags.go:64] FLAG: --register-node="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734963 4573 flags.go:64] FLAG: --register-schedulable="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734969 4573 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734977 4573 flags.go:64] FLAG: --registry-burst="10" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734983 4573 flags.go:64] FLAG: --registry-qps="5" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734987 4573 flags.go:64] FLAG: --reserved-cpus="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.734992 4573 flags.go:64] FLAG: --reserved-memory="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735001 4573 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735006 4573 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735011 4573 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735016 4573 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735021 4573 flags.go:64] FLAG: --runonce="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735025 4573 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735030 4573 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735035 4573 flags.go:64] FLAG: --seccomp-default="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735042 4573 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735066 4573 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735071 4573 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735077 4573 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735082 4573 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735203 4573 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735210 4573 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735215 4573 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735220 4573 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735226 4573 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735235 4573 flags.go:64] FLAG: --system-cgroups="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735241 4573 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735529 4573 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735534 4573 flags.go:64] FLAG: --tls-cert-file="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735539 4573 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735564 4573 flags.go:64] FLAG: --tls-min-version="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735570 4573 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735575 4573 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735579 4573 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735584 4573 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735589 4573 flags.go:64] FLAG: --v="2" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735595 4573 flags.go:64] FLAG: --version="false" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735603 4573 flags.go:64] FLAG: --vmodule="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735627 4573 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.735632 4573 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735791 4573 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735799 4573 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735804 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735809 4573 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735814 4573 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735820 4573 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735825 4573 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735829 4573 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735834 4573 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735838 4573 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735842 4573 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735847 4573 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735851 4573 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735855 4573 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735859 4573 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735862 4573 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735866 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735872 4573 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735876 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735880 4573 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735884 4573 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735887 4573 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735891 4573 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735895 4573 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735899 4573 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735902 4573 feature_gate.go:330] unrecognized feature gate: Example Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735906 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735910 4573 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735914 4573 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735917 4573 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735921 4573 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735925 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735930 4573 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735936 4573 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735940 4573 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735943 4573 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735947 4573 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735951 4573 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735987 4573 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735993 4573 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.735996 4573 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736000 4573 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736004 4573 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736007 4573 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736011 4573 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736015 4573 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736018 4573 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736022 4573 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736026 4573 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736029 4573 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736038 4573 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736042 4573 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736065 4573 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736070 4573 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736073 4573 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736080 4573 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736084 4573 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736087 4573 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736091 4573 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736095 4573 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736098 4573 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736103 4573 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736108 4573 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736111 4573 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736116 4573 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736119 4573 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736123 4573 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736127 4573 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736130 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736134 4573 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.736138 4573 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.736336 4573 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.758254 4573 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.758294 4573 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758362 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758370 4573 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758375 4573 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758384 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758389 4573 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758393 4573 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758397 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758401 4573 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758405 4573 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758409 4573 feature_gate.go:330] unrecognized feature gate: Example Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758413 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758418 4573 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758424 4573 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758428 4573 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758432 4573 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758436 4573 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758441 4573 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758445 4573 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758448 4573 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758452 4573 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758456 4573 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758460 4573 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758464 4573 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758468 4573 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758472 4573 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758475 4573 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758479 4573 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758483 4573 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758487 4573 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758491 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758495 4573 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758499 4573 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758504 4573 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758509 4573 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758515 4573 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758522 4573 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758528 4573 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758533 4573 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758537 4573 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758541 4573 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758545 4573 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758549 4573 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758553 4573 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758556 4573 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758560 4573 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758564 4573 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758569 4573 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758574 4573 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758578 4573 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758582 4573 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758586 4573 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758590 4573 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758596 4573 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758600 4573 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758605 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758609 4573 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758615 4573 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758619 4573 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758624 4573 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758628 4573 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758632 4573 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758638 4573 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758642 4573 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758646 4573 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758649 4573 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758653 4573 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758657 4573 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758661 4573 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758665 4573 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758670 4573 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758675 4573 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.758682 4573 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758808 4573 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758816 4573 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758820 4573 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758825 4573 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758828 4573 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758832 4573 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758836 4573 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758848 4573 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758853 4573 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758857 4573 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758862 4573 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758866 4573 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758899 4573 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758903 4573 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758909 4573 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758913 4573 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758917 4573 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758922 4573 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758929 4573 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758934 4573 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758938 4573 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758943 4573 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758947 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758952 4573 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758955 4573 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758960 4573 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758964 4573 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758968 4573 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758971 4573 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758976 4573 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758980 4573 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758983 4573 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758988 4573 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758993 4573 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.758997 4573 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759001 4573 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759005 4573 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759009 4573 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759012 4573 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759016 4573 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759020 4573 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759024 4573 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759027 4573 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759031 4573 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759035 4573 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759038 4573 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759061 4573 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759066 4573 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759069 4573 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759073 4573 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759078 4573 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759082 4573 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759086 4573 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759090 4573 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759094 4573 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759098 4573 feature_gate.go:330] unrecognized feature gate: Example Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759103 4573 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759107 4573 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759111 4573 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759116 4573 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759119 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759124 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759129 4573 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759133 4573 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759137 4573 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759142 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759146 4573 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759150 4573 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759155 4573 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759159 4573 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.759164 4573 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.759172 4573 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.759614 4573 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.762440 4573 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.762549 4573 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.763165 4573 server.go:997] "Starting client certificate rotation" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.763197 4573 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.763445 4573 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-09 23:14:51.455954169 +0000 UTC Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.763698 4573 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 902h36m51.692262408s for next certificate rotation Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.799081 4573 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.802029 4573 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.825233 4573 log.go:25] "Validated CRI v1 runtime API" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.851176 4573 log.go:25] "Validated CRI v1 image API" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.853570 4573 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.869981 4573 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-08-32-49-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.870038 4573 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.885311 4573 manager.go:217] Machine: {Timestamp:2025-12-03 08:37:59.884248504 +0000 UTC m=+0.452627783 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199476736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:05fa6cec-0bc2-44ec-8da4-d86d857e3ca3 BootID:56145332-ffe2-4521-8b79-5f3a443b6d75 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:24:6b:df Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:24:6b:df Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:cc:cf:98 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:15:21:1a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:b0:ba:5a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b5:19:cd Speed:-1 Mtu:1496} {Name:eth10 MacAddress:c6:95:23:a2:b6:0d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:66:00:d6:db:c8:1a Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199476736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.885538 4573 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.885758 4573 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.886257 4573 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.886440 4573 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.886493 4573 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.886727 4573 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.886739 4573 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.886995 4573 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.887038 4573 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.887384 4573 state_mem.go:36] "Initialized new in-memory state store" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.887484 4573 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.888152 4573 kubelet.go:418] "Attempting to sync node with API server" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.888176 4573 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.888204 4573 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.888222 4573 kubelet.go:324] "Adding apiserver pod source" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.888236 4573 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.891070 4573 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.891512 4573 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.892256 4573 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.892661 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:37:59 crc kubenswrapper[4573]: E1203 08:37:59.892836 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.892939 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.892961 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.892970 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.892978 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.892991 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.892998 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893006 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893017 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893026 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893036 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893065 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893074 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893300 4573 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.893678 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:37:59 crc kubenswrapper[4573]: E1203 08:37:59.893901 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893888 4573 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.894213 4573 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.893856 4573 server.go:1280] "Started kubelet" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.896017 4573 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.896959 4573 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.897006 4573 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 08:37:59 crc kubenswrapper[4573]: E1203 08:37:59.896112 4573 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187da7c1d20f01d7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 08:37:59.893828055 +0000 UTC m=+0.462207304,LastTimestamp:2025-12-03 08:37:59.893828055 +0000 UTC m=+0.462207304,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.898936 4573 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 06:49:49.338268156 +0000 UTC Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.899007 4573 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1006h11m49.439269025s for next certificate rotation Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.899261 4573 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.899328 4573 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.899499 4573 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 08:37:59 crc kubenswrapper[4573]: E1203 08:37:59.900702 4573 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 08:37:59 crc kubenswrapper[4573]: E1203 08:37:59.900888 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="200ms" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.906874 4573 factory.go:55] Registering systemd factory Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.906944 4573 factory.go:221] Registration of the systemd container factory successfully Dec 03 08:37:59 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.920202 4573 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:37:59 crc kubenswrapper[4573]: W1203 08:37:59.920910 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:37:59 crc kubenswrapper[4573]: E1203 08:37:59.921011 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.921333 4573 factory.go:153] Registering CRI-O factory Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.921636 4573 factory.go:221] Registration of the crio container factory successfully Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.921738 4573 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.921772 4573 factory.go:103] Registering Raw factory Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.921805 4573 manager.go:1196] Started watching for new ooms in manager Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.922503 4573 server.go:460] "Adding debug handlers to kubelet server" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.925034 4573 manager.go:319] Starting recovery of all containers Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.960847 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.960968 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.960992 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961023 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961072 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961097 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961118 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961143 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961173 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961196 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961222 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961243 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961268 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961325 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961347 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961368 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961524 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961552 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961606 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961631 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961654 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961676 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961699 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961724 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961747 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961770 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961801 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961823 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961850 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961871 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961891 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961911 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961933 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961954 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961975 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.961995 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962014 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962032 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962104 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962123 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962144 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962161 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962180 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962197 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962219 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962235 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962256 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962279 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962299 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962318 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962335 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962352 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962374 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962392 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962411 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962430 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962449 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962469 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962486 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962504 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962526 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962547 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962568 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962591 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962612 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962632 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962652 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.962674 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.965845 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.965890 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.965912 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.965929 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.965949 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.965969 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.965988 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966006 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966027 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966086 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966104 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966122 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966142 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966160 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966179 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966199 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966218 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966242 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966259 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966276 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966296 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966316 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966330 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966344 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966361 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966378 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966392 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966408 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966423 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966439 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966457 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966472 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966486 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966505 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966521 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966537 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.966561 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967288 4573 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967342 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967361 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967382 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967397 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967411 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967427 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967441 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967457 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967474 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967491 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967505 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967520 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967534 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967547 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967562 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967577 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967591 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967605 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967623 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967638 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967651 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967666 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967682 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967697 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967710 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967724 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967738 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967751 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967765 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967779 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967794 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967811 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967825 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967840 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967856 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967870 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967884 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967899 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967936 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967951 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967969 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967984 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.967998 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968013 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968027 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968042 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968093 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968107 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968120 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968135 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968150 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968164 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968178 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968193 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968209 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968225 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968240 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968254 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968268 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968283 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968297 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968313 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968327 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968340 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968354 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968370 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968384 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968412 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968426 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968439 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968452 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968470 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968484 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968498 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968511 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968526 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968541 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968554 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968569 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968582 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968598 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968613 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968628 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968643 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968659 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968681 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968701 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968730 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968750 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968770 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968788 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968816 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968836 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968854 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968878 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968896 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968915 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968936 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968958 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968978 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.968995 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.969011 4573 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.969027 4573 reconstruct.go:97] "Volume reconstruction finished" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.969042 4573 reconciler.go:26] "Reconciler: start to sync state" Dec 03 08:37:59 crc kubenswrapper[4573]: I1203 08:37:59.984751 4573 manager.go:324] Recovery completed Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:37:59.999975 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.001114 4573 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.002390 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.002646 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.002658 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.003997 4573 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.004016 4573 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.004057 4573 state_mem.go:36] "Initialized new in-memory state store" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.013961 4573 policy_none.go:49] "None policy: Start" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.015810 4573 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.015860 4573 state_mem.go:35] "Initializing new in-memory state store" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.027302 4573 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.029027 4573 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.029115 4573 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.029143 4573 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.029194 4573 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.030332 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.030416 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.075531 4573 manager.go:334] "Starting Device Plugin manager" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.075594 4573 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.075607 4573 server.go:79] "Starting device plugin registration server" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.076011 4573 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.076033 4573 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.076346 4573 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.076513 4573 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.076531 4573 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.090104 4573 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.101833 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="400ms" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.130144 4573 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.130326 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.131422 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.131460 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.131471 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.131636 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.132032 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.132203 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.132393 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.132438 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.132450 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.132660 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.132778 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.132855 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.133530 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.133555 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.133565 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.133688 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.133913 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.133962 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134234 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134287 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134303 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134330 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134356 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134366 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134499 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134629 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.134673 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135133 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135163 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135171 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135340 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135372 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135628 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135692 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135708 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135809 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.135900 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.136017 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.136070 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.136230 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.136246 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.136106 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.136289 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.136299 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.177180 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.179256 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.179295 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.179309 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.179341 4573 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.179750 4573 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.273753 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.273819 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274405 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274449 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274477 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274516 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274540 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274558 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274578 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274602 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274622 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274640 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274659 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274679 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.274709 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376510 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376737 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376814 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376850 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376986 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377006 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377055 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377079 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377156 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376897 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376957 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377213 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377256 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377280 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377283 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376924 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377332 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377353 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377355 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.376932 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377332 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377406 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377447 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377474 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377499 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377518 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377473 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377500 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377511 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.377597 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.379987 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.383576 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.383621 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.383634 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.383668 4573 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.384236 4573 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.458478 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.476497 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.482885 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.492955 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5df27bfc5a0998dc011f209e4a8491ae981902a9711f3d692e9480ed137f7a86 WatchSource:0}: Error finding container 5df27bfc5a0998dc011f209e4a8491ae981902a9711f3d692e9480ed137f7a86: Status 404 returned error can't find the container with id 5df27bfc5a0998dc011f209e4a8491ae981902a9711f3d692e9480ed137f7a86 Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.503956 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.503976 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="800ms" Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.505869 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-737ad0deca618a45973732d392ab2a2206a057f31fe5ee8300871147d9f01e90 WatchSource:0}: Error finding container 737ad0deca618a45973732d392ab2a2206a057f31fe5ee8300871147d9f01e90: Status 404 returned error can't find the container with id 737ad0deca618a45973732d392ab2a2206a057f31fe5ee8300871147d9f01e90 Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.510483 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.514165 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-9a98c9d6f5f980a409d705472dc437f167956fddbc91c42e394911e592e8e2f8 WatchSource:0}: Error finding container 9a98c9d6f5f980a409d705472dc437f167956fddbc91c42e394911e592e8e2f8: Status 404 returned error can't find the container with id 9a98c9d6f5f980a409d705472dc437f167956fddbc91c42e394911e592e8e2f8 Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.527141 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-eb4726572a39dd0274f2c59f76fe521be88edd109f27f14df073b9dadc89d36b WatchSource:0}: Error finding container eb4726572a39dd0274f2c59f76fe521be88edd109f27f14df073b9dadc89d36b: Status 404 returned error can't find the container with id eb4726572a39dd0274f2c59f76fe521be88edd109f27f14df073b9dadc89d36b Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.532513 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ff342f812f7500f0850bd62fc5741b549996afb68d099fbad5bd9923411b85fc WatchSource:0}: Error finding container ff342f812f7500f0850bd62fc5741b549996afb68d099fbad5bd9923411b85fc: Status 404 returned error can't find the container with id ff342f812f7500f0850bd62fc5741b549996afb68d099fbad5bd9923411b85fc Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.757722 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.757872 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.784401 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.785567 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.785608 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.785622 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.785648 4573 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.785990 4573 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Dec 03 08:38:00 crc kubenswrapper[4573]: I1203 08:38:00.922835 4573 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.976964 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.977072 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:38:00 crc kubenswrapper[4573]: W1203 08:38:00.986628 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:00 crc kubenswrapper[4573]: E1203 08:38:00.986754 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.036567 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6ee9b66a1bc72eda0261abd5ee7d063582d9ac4127c87dd20f5c3a5684ca8cc9"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.036756 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.037939 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.038011 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.038025 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.036484 4573 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6ee9b66a1bc72eda0261abd5ee7d063582d9ac4127c87dd20f5c3a5684ca8cc9" exitCode=0 Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.039072 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"737ad0deca618a45973732d392ab2a2206a057f31fe5ee8300871147d9f01e90"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.042590 4573 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="61424a4f23000a56999960cbdb89e67e37b0bd43a74219bcf88fb6a50dff1da4" exitCode=0 Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.042696 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"61424a4f23000a56999960cbdb89e67e37b0bd43a74219bcf88fb6a50dff1da4"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.042744 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5df27bfc5a0998dc011f209e4a8491ae981902a9711f3d692e9480ed137f7a86"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.042883 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.043994 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.044029 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.044039 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.045212 4573 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3" exitCode=0 Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.045314 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.045343 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ff342f812f7500f0850bd62fc5741b549996afb68d099fbad5bd9923411b85fc"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.045432 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.046165 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.046186 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.046195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.047841 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.047881 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eb4726572a39dd0274f2c59f76fe521be88edd109f27f14df073b9dadc89d36b"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.049738 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966" exitCode=0 Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.049776 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.049808 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9a98c9d6f5f980a409d705472dc437f167956fddbc91c42e394911e592e8e2f8"} Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.049935 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.050684 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.050703 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.050712 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.052956 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.053749 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.053793 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.053812 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:01 crc kubenswrapper[4573]: E1203 08:38:01.305305 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="1.6s" Dec 03 08:38:01 crc kubenswrapper[4573]: W1203 08:38:01.570883 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:01 crc kubenswrapper[4573]: E1203 08:38:01.570993 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.586154 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.587731 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.587787 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.587801 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.587844 4573 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 08:38:01 crc kubenswrapper[4573]: E1203 08:38:01.588405 4573 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Dec 03 08:38:01 crc kubenswrapper[4573]: I1203 08:38:01.957235 4573 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.060293 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2773e32404b31160beec1d08cef6235ae7c2909d701b7fbcfce92152b2d4fafe"} Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.060390 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.061396 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.061421 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.061431 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.063113 4573 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2d21665c9762f4c5794e20734889f39231981a2c0c677e8090ad2980e85e1755" exitCode=0 Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.063162 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2d21665c9762f4c5794e20734889f39231981a2c0c677e8090ad2980e85e1755"} Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.063244 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.064231 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.064254 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.064265 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.068411 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82"} Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.068459 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f"} Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.071793 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab"} Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.071818 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3"} Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.075002 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97"} Dec 03 08:38:02 crc kubenswrapper[4573]: W1203 08:38:02.510431 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:02 crc kubenswrapper[4573]: E1203 08:38:02.510583 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:38:02 crc kubenswrapper[4573]: E1203 08:38:02.912191 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="3.2s" Dec 03 08:38:02 crc kubenswrapper[4573]: I1203 08:38:02.920872 4573 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.086803 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4"} Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.086845 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9"} Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.086854 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57"} Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.092242 4573 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="87742c98718b8ff664bb42a7abb281b1465df3e245c35b941ec825fb396547f1" exitCode=0 Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.092292 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"87742c98718b8ff664bb42a7abb281b1465df3e245c35b941ec825fb396547f1"} Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.092408 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.093364 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.093380 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.093389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.096391 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea"} Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.096429 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.098465 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.098505 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.098512 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.108659 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.109256 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.109615 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c"} Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.110397 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.110422 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.110432 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.110949 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.110965 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.110974 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.189515 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.190411 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.190439 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.190447 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.190467 4573 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 08:38:03 crc kubenswrapper[4573]: E1203 08:38:03.190842 4573 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Dec 03 08:38:03 crc kubenswrapper[4573]: W1203 08:38:03.363846 4573 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Dec 03 08:38:03 crc kubenswrapper[4573]: E1203 08:38:03.363943 4573 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Dec 03 08:38:03 crc kubenswrapper[4573]: I1203 08:38:03.642498 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.117309 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"58c67c54dbacf5ef99060c0aa0fc9afb05148989645cb74acb236a3107b3d9d3"} Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.117384 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d4ea1ae440959c88c05edb78b0ef1031186e899821b71e589a2e91776f355f2e"} Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.117397 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6abdb217c16638eac968f2ca5423deacf945c63f23308388f50b6a7842dcbf27"} Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.117408 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b7365e4e47cd6dc02ccee8b96fadedc0a9779682191b812efe58606f4d371a82"} Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.121568 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05"} Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.121726 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.121767 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.121787 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.121833 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123316 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123348 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123378 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123350 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123439 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123321 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123527 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.123537 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:04 crc kubenswrapper[4573]: I1203 08:38:04.875313 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.129809 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.129913 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.130554 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.130562 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a686c18023ff96fec4885ee571151253ff025601b7afb973b13c54885dff1cca"} Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.130765 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.130924 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.130975 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.130998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.131079 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.131115 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.131125 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.132105 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.132163 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:05 crc kubenswrapper[4573]: I1203 08:38:05.132186 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.133458 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.133540 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.135270 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.135542 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.135743 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.135458 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.135985 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.136005 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.391758 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.393715 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.393762 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.393774 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.393805 4573 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.643187 4573 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.643785 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.968365 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.968641 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.970412 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.970459 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:06 crc kubenswrapper[4573]: I1203 08:38:06.970482 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:07 crc kubenswrapper[4573]: I1203 08:38:07.137152 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:07 crc kubenswrapper[4573]: I1203 08:38:07.137348 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:07 crc kubenswrapper[4573]: I1203 08:38:07.138344 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:07 crc kubenswrapper[4573]: I1203 08:38:07.138378 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:07 crc kubenswrapper[4573]: I1203 08:38:07.138389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:08 crc kubenswrapper[4573]: I1203 08:38:08.092215 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:08 crc kubenswrapper[4573]: I1203 08:38:08.092455 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:08 crc kubenswrapper[4573]: I1203 08:38:08.093989 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:08 crc kubenswrapper[4573]: I1203 08:38:08.094127 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:08 crc kubenswrapper[4573]: I1203 08:38:08.094152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:09 crc kubenswrapper[4573]: I1203 08:38:09.540123 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:09 crc kubenswrapper[4573]: I1203 08:38:09.540374 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:09 crc kubenswrapper[4573]: I1203 08:38:09.542126 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:09 crc kubenswrapper[4573]: I1203 08:38:09.542200 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:09 crc kubenswrapper[4573]: I1203 08:38:09.542228 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:10 crc kubenswrapper[4573]: E1203 08:38:10.090298 4573 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.100656 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.100920 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.102568 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.102636 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.102670 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.708730 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.709033 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.711182 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.711268 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.711292 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:10 crc kubenswrapper[4573]: I1203 08:38:10.716343 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:11 crc kubenswrapper[4573]: I1203 08:38:11.148539 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:11 crc kubenswrapper[4573]: I1203 08:38:11.149901 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:11 crc kubenswrapper[4573]: I1203 08:38:11.149961 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:11 crc kubenswrapper[4573]: I1203 08:38:11.149982 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:11 crc kubenswrapper[4573]: I1203 08:38:11.153872 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:12 crc kubenswrapper[4573]: I1203 08:38:12.151588 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:12 crc kubenswrapper[4573]: I1203 08:38:12.152967 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:12 crc kubenswrapper[4573]: I1203 08:38:12.153098 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:12 crc kubenswrapper[4573]: I1203 08:38:12.153122 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:13 crc kubenswrapper[4573]: I1203 08:38:13.915031 4573 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 08:38:13 crc kubenswrapper[4573]: I1203 08:38:13.915100 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 08:38:13 crc kubenswrapper[4573]: I1203 08:38:13.923173 4573 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 08:38:13 crc kubenswrapper[4573]: I1203 08:38:13.923219 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 08:38:14 crc kubenswrapper[4573]: I1203 08:38:14.887677 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 08:38:14 crc kubenswrapper[4573]: I1203 08:38:14.887865 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:14 crc kubenswrapper[4573]: I1203 08:38:14.889302 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:14 crc kubenswrapper[4573]: I1203 08:38:14.889356 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:14 crc kubenswrapper[4573]: I1203 08:38:14.889369 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:14 crc kubenswrapper[4573]: I1203 08:38:14.927240 4573 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]log ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]etcd ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/priority-and-fairness-filter ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-apiextensions-informers ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-apiextensions-controllers ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/crd-informer-synced ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-system-namespaces-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 03 08:38:14 crc kubenswrapper[4573]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 03 08:38:14 crc kubenswrapper[4573]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/bootstrap-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/start-kube-aggregator-informers ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/apiservice-registration-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/apiservice-discovery-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]autoregister-completion ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/apiservice-openapi-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 03 08:38:14 crc kubenswrapper[4573]: livez check failed Dec 03 08:38:14 crc kubenswrapper[4573]: I1203 08:38:14.927307 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:38:15 crc kubenswrapper[4573]: I1203 08:38:15.611665 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 08:38:15 crc kubenswrapper[4573]: I1203 08:38:15.611888 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:15 crc kubenswrapper[4573]: I1203 08:38:15.613606 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:15 crc kubenswrapper[4573]: I1203 08:38:15.613669 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:15 crc kubenswrapper[4573]: I1203 08:38:15.613688 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:15 crc kubenswrapper[4573]: I1203 08:38:15.626037 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 08:38:16 crc kubenswrapper[4573]: I1203 08:38:16.161712 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:16 crc kubenswrapper[4573]: I1203 08:38:16.163191 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:16 crc kubenswrapper[4573]: I1203 08:38:16.163226 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:16 crc kubenswrapper[4573]: I1203 08:38:16.163237 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:16 crc kubenswrapper[4573]: I1203 08:38:16.643138 4573 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 08:38:16 crc kubenswrapper[4573]: I1203 08:38:16.643715 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 08:38:18 crc kubenswrapper[4573]: E1203 08:38:18.905228 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.910856 4573 trace.go:236] Trace[89402780]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 08:38:04.046) (total time: 14864ms): Dec 03 08:38:18 crc kubenswrapper[4573]: Trace[89402780]: ---"Objects listed" error: 14864ms (08:38:18.910) Dec 03 08:38:18 crc kubenswrapper[4573]: Trace[89402780]: [14.86448846s] [14.86448846s] END Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.910896 4573 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.911427 4573 trace.go:236] Trace[1195749107]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 08:38:07.956) (total time: 10954ms): Dec 03 08:38:18 crc kubenswrapper[4573]: Trace[1195749107]: ---"Objects listed" error: 10954ms (08:38:18.911) Dec 03 08:38:18 crc kubenswrapper[4573]: Trace[1195749107]: [10.954974459s] [10.954974459s] END Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.911443 4573 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.911800 4573 trace.go:236] Trace[1985415424]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 08:38:04.027) (total time: 14884ms): Dec 03 08:38:18 crc kubenswrapper[4573]: Trace[1985415424]: ---"Objects listed" error: 14884ms (08:38:18.911) Dec 03 08:38:18 crc kubenswrapper[4573]: Trace[1985415424]: [14.884198155s] [14.884198155s] END Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.911828 4573 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.912456 4573 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.912713 4573 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 08:38:18 crc kubenswrapper[4573]: E1203 08:38:18.924757 4573 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.953719 4573 apiserver.go:52] "Watching apiserver" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.957008 4573 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.957286 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.958226 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:18 crc kubenswrapper[4573]: E1203 08:38:18.958318 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.958404 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.958782 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:18 crc kubenswrapper[4573]: E1203 08:38:18.958827 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.959113 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:18 crc kubenswrapper[4573]: E1203 08:38:18.959185 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.959271 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.959608 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.961736 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.961783 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.962766 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.962845 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.962957 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.965592 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.966526 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.967031 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 08:38:18 crc kubenswrapper[4573]: I1203 08:38:18.969179 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.009296 4573 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.013903 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.013953 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.013981 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014006 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014026 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014066 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014090 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014114 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014134 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014155 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014176 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014197 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014221 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014244 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014455 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014481 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014526 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014552 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014577 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014602 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014608 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014626 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014651 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014677 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014702 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014727 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014754 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014775 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014799 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014809 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014821 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014841 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014899 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014924 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014945 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.014979 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015004 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015029 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015039 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015074 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015099 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015121 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015142 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015163 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015176 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015189 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015226 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015246 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015268 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015289 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015308 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015326 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015325 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015345 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015365 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015384 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015404 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015421 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015438 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015455 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015471 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015488 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015506 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015523 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015537 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015552 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015567 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015584 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015593 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015603 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015747 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015826 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015863 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015888 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015914 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015941 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015966 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.015990 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016015 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016039 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016085 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016108 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016164 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016169 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016189 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016211 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016232 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016256 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016279 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016299 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016319 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016342 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016376 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016399 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016423 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016447 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016469 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016492 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016514 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016537 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016559 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016581 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016606 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016662 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016684 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016706 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016749 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016791 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016813 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016850 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016874 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016897 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016932 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016957 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.017018 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.017061 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.017095 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.017116 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.020887 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024360 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024402 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024435 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024468 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024496 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024529 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024556 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024583 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024678 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024708 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024736 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024766 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024791 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024817 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024843 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024882 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024906 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016234 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016373 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.016567 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.017729 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.017809 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.017844 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018011 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018175 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018220 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018302 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018479 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018560 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018717 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038818 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018775 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.018983 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.019303 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.019313 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.019490 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.019487 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.019786 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038903 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.019882 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.019987 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.020242 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.020470 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.020781 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.021734 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.022261 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.022470 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.025270 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.025456 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.025760 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.025869 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.025939 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.026149 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.026321 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.026507 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.026585 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.026751 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.026848 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.026910 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.027285 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.027609 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.027830 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.028702 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.028958 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.039291 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.029277 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.029555 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.029819 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.029833 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.030130 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.030585 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.030691 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.031025 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.031462 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.031482 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.031763 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.031984 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032207 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032281 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.039435 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032358 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032429 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032458 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032642 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032718 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032832 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032931 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032969 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.033101 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.033324 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.033810 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.034198 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.034228 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.034369 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:38:19.534346373 +0000 UTC m=+20.102725632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.039869 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.040275 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.040348 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.034523 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.035319 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.034331 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.035629 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.035689 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.035975 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.036001 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.035992 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.036381 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.036467 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.036541 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.036619 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.037190 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.037286 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.037910 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038083 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038104 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.040627 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038283 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038508 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038523 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038759 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.038779 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.039085 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.039135 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.039200 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.032333 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.039527 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.041432 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.041522 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.041822 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.024931 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.047667 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.047894 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.047935 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.047967 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.047995 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048027 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048073 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048101 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048129 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048156 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048194 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048222 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048247 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048272 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048298 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048324 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048349 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048373 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048397 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048422 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048443 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048465 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048488 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048510 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048531 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048554 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048576 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048596 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048618 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048639 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048661 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048684 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048712 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048732 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048754 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.048888 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.049917 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.058252 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.062908 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063023 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063074 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063109 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063134 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063163 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063186 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063212 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063236 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063262 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063294 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063322 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063343 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063368 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063394 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063419 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063452 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063483 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063509 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063534 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063564 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063591 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063618 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063648 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.063670 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064120 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064165 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064195 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064220 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064328 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064373 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064402 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064429 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064459 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064489 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064524 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064558 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064586 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064613 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064642 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064669 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064693 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064721 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064864 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064890 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064904 4573 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064947 4573 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064962 4573 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064975 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.064989 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065002 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065017 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065031 4573 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065063 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065079 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065092 4573 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065105 4573 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065119 4573 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065132 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065145 4573 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065158 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065172 4573 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065186 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.065199 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066139 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066285 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066308 4573 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066323 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066338 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066353 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066370 4573 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066384 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066400 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066415 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066429 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066444 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066459 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066472 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066485 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066498 4573 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066510 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066523 4573 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066536 4573 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066549 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066560 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066573 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.066585 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.067822 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.072173 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.072365 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.072541 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.072722 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.072870 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.077340 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.079422 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.079843 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.080195 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.080797 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.081072 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.081726 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.084159 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.084298 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.084517 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.084764 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.084794 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.085092 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.085152 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.085581 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.085674 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.085992 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.086140 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.087350 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.087843 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.088016 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.088392 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.088999 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.089147 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.089346 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.089520 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.089673 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.089877 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.089884 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.096436 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.097008 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.097034 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.097127 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.097256 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.100303 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.097392 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.100410 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.100750 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.100762 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.101187 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.101462 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.103138 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.099214 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.103586 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.103789 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.104152 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.104876 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.104912 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.107560 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.107601 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.107954 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.108040 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.108361 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.108691 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.109086 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.109613 4573 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.109722 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:19.609692001 +0000 UTC m=+20.178071470 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.109885 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.110271 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.110337 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.111631 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.112121 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.113133 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.113145 4573 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.114401 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.116210 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.118224 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.120582 4573 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.120698 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:19.620670409 +0000 UTC m=+20.189049828 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.120973 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121362 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121450 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121466 4573 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121526 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121543 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121556 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121576 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121590 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121606 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121617 4573 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121630 4573 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121643 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121659 4573 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121673 4573 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121689 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121703 4573 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121716 4573 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121728 4573 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121741 4573 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121754 4573 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121766 4573 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121778 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121793 4573 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121805 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121820 4573 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121848 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121860 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121872 4573 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121886 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121898 4573 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121911 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121921 4573 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121931 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121941 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121952 4573 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121962 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121973 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121982 4573 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.121992 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.122002 4573 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.122013 4573 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.122022 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.122032 4573 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.122042 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143162 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.126944 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.128462 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.129472 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.138358 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.143334 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.143356 4573 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.143434 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:19.643413524 +0000 UTC m=+20.211792783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143174 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143481 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143493 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143503 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143513 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143525 4573 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143538 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143549 4573 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143559 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143571 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143581 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143593 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143604 4573 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143616 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143626 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143637 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143647 4573 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143658 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143668 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143678 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143689 4573 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143700 4573 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143713 4573 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143723 4573 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143733 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143744 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143754 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143765 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143775 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143786 4573 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143797 4573 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143808 4573 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.143818 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.123237 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.144368 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.155779 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.162492 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.163961 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.172766 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.179175 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.179339 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.179419 4573 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.179597 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:19.679564613 +0000 UTC m=+20.247943872 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.198173 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.202829 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.227396 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.244980 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245037 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245102 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245114 4573 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245123 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245132 4573 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245141 4573 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245155 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245166 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245177 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245190 4573 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245201 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245213 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245238 4573 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245250 4573 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245261 4573 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245272 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245281 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245290 4573 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245298 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245312 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245323 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245337 4573 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245348 4573 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245360 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245372 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245384 4573 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245395 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245403 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245411 4573 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245420 4573 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245428 4573 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245436 4573 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245445 4573 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245454 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245463 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245473 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245481 4573 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245489 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245498 4573 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245506 4573 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245514 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245523 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245532 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245540 4573 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245548 4573 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245557 4573 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245565 4573 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245573 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245582 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245591 4573 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245600 4573 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245611 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245619 4573 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245627 4573 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245636 4573 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245646 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245655 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245668 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245682 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245693 4573 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245703 4573 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245713 4573 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245721 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245730 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245738 4573 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245746 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245755 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245763 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245771 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245779 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245788 4573 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245796 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245804 4573 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245811 4573 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245821 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245831 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.245953 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.246001 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.283267 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.291701 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.296749 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.299638 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.330452 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: W1203 08:38:19.338428 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-1d909c66a1be38c54b1590de87f1746332fc8bd1272c554316793e91a649fb05 WatchSource:0}: Error finding container 1d909c66a1be38c54b1590de87f1746332fc8bd1272c554316793e91a649fb05: Status 404 returned error can't find the container with id 1d909c66a1be38c54b1590de87f1746332fc8bd1272c554316793e91a649fb05 Dec 03 08:38:19 crc kubenswrapper[4573]: W1203 08:38:19.346874 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-a10c8de206fa717aaa099e6dc7dd448753e65083b6ee2d07fb7f412e2642cbfc WatchSource:0}: Error finding container a10c8de206fa717aaa099e6dc7dd448753e65083b6ee2d07fb7f412e2642cbfc: Status 404 returned error can't find the container with id a10c8de206fa717aaa099e6dc7dd448753e65083b6ee2d07fb7f412e2642cbfc Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.430591 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.457453 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.479510 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.497399 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-xg4ms"] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.497800 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-zxsgx"] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.497959 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-ch24s"] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.498240 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.498611 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.499003 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zxsgx" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.507247 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.506622 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.593291 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.593573 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.594200 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.595909 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.605095 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.605621 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.606440 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-mfcvl"] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.617233 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.617833 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.617997 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618152 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618332 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618585 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618710 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89drg\" (UniqueName: \"kubernetes.io/projected/b73da931-0de5-4fd1-b8ff-878cc7008bf3-kube-api-access-89drg\") pod \"node-resolver-zxsgx\" (UID: \"b73da931-0de5-4fd1-b8ff-878cc7008bf3\") " pod="openshift-dns/node-resolver-zxsgx" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618770 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-conf-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618798 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b73da931-0de5-4fd1-b8ff-878cc7008bf3-hosts-file\") pod \"node-resolver-zxsgx\" (UID: \"b73da931-0de5-4fd1-b8ff-878cc7008bf3\") " pod="openshift-dns/node-resolver-zxsgx" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618829 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-cnibin\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618854 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-socket-dir-parent\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618884 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e17325e8-2a46-4028-b7c5-664eeda96013-proxy-tls\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618914 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-hostroot\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618941 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-os-release\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618966 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-cni-binary-copy\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.618989 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-system-cni-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619013 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-cni-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619038 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-cni-multus\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619083 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-daemon-config\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619118 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619144 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e17325e8-2a46-4028-b7c5-664eeda96013-rootfs\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619169 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e17325e8-2a46-4028-b7c5-664eeda96013-mcd-auth-proxy-config\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619198 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b2qt\" (UniqueName: \"kubernetes.io/projected/e17325e8-2a46-4028-b7c5-664eeda96013-kube-api-access-7b2qt\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619222 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-multus-certs\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619261 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-etc-kubernetes\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619284 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srgkq\" (UniqueName: \"kubernetes.io/projected/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-kube-api-access-srgkq\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619306 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-k8s-cni-cncf-io\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619330 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-netns\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619363 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-cni-bin\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619385 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-kubelet\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.619576 4573 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.619660 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:20.619640047 +0000 UTC m=+21.188019306 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.619680 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:38:20.619672249 +0000 UTC m=+21.188051508 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.619880 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.620090 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.625023 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.625318 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.634874 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.651621 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.687933 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.708642 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727147 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e17325e8-2a46-4028-b7c5-664eeda96013-proxy-tls\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727208 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-cnibin\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727251 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-os-release\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727287 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-cni-binary-copy\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727318 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e17325e8-2a46-4028-b7c5-664eeda96013-rootfs\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727408 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-system-cni-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727467 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-cni-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727516 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b2qt\" (UniqueName: \"kubernetes.io/projected/e17325e8-2a46-4028-b7c5-664eeda96013-kube-api-access-7b2qt\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727547 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-multus-certs\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727541 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-cnibin\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727574 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.727741 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.727773 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727775 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-kubelet\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.727793 4573 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727816 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-netns\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727856 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-conf-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.727872 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:20.727852577 +0000 UTC m=+21.296231836 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727905 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89drg\" (UniqueName: \"kubernetes.io/projected/b73da931-0de5-4fd1-b8ff-878cc7008bf3-kube-api-access-89drg\") pod \"node-resolver-zxsgx\" (UID: \"b73da931-0de5-4fd1-b8ff-878cc7008bf3\") " pod="openshift-dns/node-resolver-zxsgx" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727945 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-kubelet\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727909 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-conf-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.727945 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.728142 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e17325e8-2a46-4028-b7c5-664eeda96013-rootfs\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.728191 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-system-cni-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.728550 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-cni-dir\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.728878 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-multus-certs\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729083 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-netns\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729110 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729138 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729142 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-cni-binary-copy\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729168 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.729209 4573 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729251 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-socket-dir-parent\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729305 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cnibin\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729421 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-socket-dir-parent\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729471 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-system-cni-dir\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.729569 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:20.729545143 +0000 UTC m=+21.297924402 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729625 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-os-release\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729658 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-hostroot\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729912 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-os-release\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729952 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-hostroot\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.729994 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-daemon-config\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.730097 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e17325e8-2a46-4028-b7c5-664eeda96013-mcd-auth-proxy-config\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.730125 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-cni-multus\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.730185 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-etc-kubernetes\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.730218 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srgkq\" (UniqueName: \"kubernetes.io/projected/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-kube-api-access-srgkq\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.730244 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj9xr\" (UniqueName: \"kubernetes.io/projected/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-kube-api-access-pj9xr\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.730278 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-cni-multus\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.730994 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-multus-daemon-config\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.733352 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-etc-kubernetes\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.733477 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-k8s-cni-cncf-io\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.733537 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-cni-bin\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.733583 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.733648 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b73da931-0de5-4fd1-b8ff-878cc7008bf3-hosts-file\") pod \"node-resolver-zxsgx\" (UID: \"b73da931-0de5-4fd1-b8ff-878cc7008bf3\") " pod="openshift-dns/node-resolver-zxsgx" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.733724 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b73da931-0de5-4fd1-b8ff-878cc7008bf3-hosts-file\") pod \"node-resolver-zxsgx\" (UID: \"b73da931-0de5-4fd1-b8ff-878cc7008bf3\") " pod="openshift-dns/node-resolver-zxsgx" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.733749 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-run-k8s-cni-cncf-io\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.733810 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-host-var-lib-cni-bin\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.733868 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.733910 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.733920 4573 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:19 crc kubenswrapper[4573]: E1203 08:38:19.733952 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:20.733940692 +0000 UTC m=+21.302319951 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.741028 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e17325e8-2a46-4028-b7c5-664eeda96013-mcd-auth-proxy-config\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.750169 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e17325e8-2a46-4028-b7c5-664eeda96013-proxy-tls\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.750438 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.781740 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srgkq\" (UniqueName: \"kubernetes.io/projected/8c98ef96-0e91-49e3-88d5-c95dd6f52c5d-kube-api-access-srgkq\") pod \"multus-ch24s\" (UID: \"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\") " pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.784495 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b2qt\" (UniqueName: \"kubernetes.io/projected/e17325e8-2a46-4028-b7c5-664eeda96013-kube-api-access-7b2qt\") pod \"machine-config-daemon-xg4ms\" (UID: \"e17325e8-2a46-4028-b7c5-664eeda96013\") " pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.808635 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89drg\" (UniqueName: \"kubernetes.io/projected/b73da931-0de5-4fd1-b8ff-878cc7008bf3-kube-api-access-89drg\") pod \"node-resolver-zxsgx\" (UID: \"b73da931-0de5-4fd1-b8ff-878cc7008bf3\") " pod="openshift-dns/node-resolver-zxsgx" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.817790 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.834636 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.835492 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.835526 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.835550 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cnibin\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.835576 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-system-cni-dir\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.835608 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj9xr\" (UniqueName: \"kubernetes.io/projected/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-kube-api-access-pj9xr\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.835650 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-os-release\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.835733 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-os-release\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.835443 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.836315 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.836895 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-cnibin\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.836953 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-system-cni-dir\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.837037 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.855850 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.860704 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj9xr\" (UniqueName: \"kubernetes.io/projected/60e19539-4df1-4909-b8ee-c6ac0fcfc52d-kube-api-access-pj9xr\") pod \"multus-additional-cni-plugins-mfcvl\" (UID: \"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\") " pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.873708 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.884952 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.885660 4573 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.885740 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.891993 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.898853 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.920403 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.920774 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x466z"] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.921892 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-blvnm"] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.924114 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.926520 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.929020 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.929854 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.931141 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.933197 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.938521 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.938725 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.938766 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.939002 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.939181 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.939296 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.939415 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.941077 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.945395 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.964772 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.967448 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-ch24s" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.976147 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.990112 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zxsgx" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.990250 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:19 crc kubenswrapper[4573]: I1203 08:38:19.993357 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.034151 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.034559 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.034835 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.049776 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fssxg\" (UniqueName: \"kubernetes.io/projected/48186949-d112-4fb7-aa60-026d0176c73e-kube-api-access-fssxg\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.049827 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-netd\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.049850 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-ovn-kubernetes\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.049970 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-kubelet\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050033 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-systemd-units\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050115 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-ovn\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050142 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48186949-d112-4fb7-aa60-026d0176c73e-ovn-node-metrics-cert\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050170 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-var-lib-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050212 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56192d0e-1b05-4a3d-b176-b3d27de36527-host\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050252 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050275 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-env-overrides\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050293 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-etc-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050310 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-node-log\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050345 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-slash\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050365 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-log-socket\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050394 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050414 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56192d0e-1b05-4a3d-b176-b3d27de36527-serviceca\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050432 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-netns\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050450 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-config\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050471 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsm9q\" (UniqueName: \"kubernetes.io/projected/56192d0e-1b05-4a3d-b176-b3d27de36527-kube-api-access-vsm9q\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050507 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-script-lib\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050592 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-systemd\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.050635 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-bin\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.070438 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: W1203 08:38:20.071181 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60e19539_4df1_4909_b8ee_c6ac0fcfc52d.slice/crio-cbceac120f35446082c02c5a52c760e14c71f8bd806460a5cf7c21f38a109cfd WatchSource:0}: Error finding container cbceac120f35446082c02c5a52c760e14c71f8bd806460a5cf7c21f38a109cfd: Status 404 returned error can't find the container with id cbceac120f35446082c02c5a52c760e14c71f8bd806460a5cf7c21f38a109cfd Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.074416 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.075393 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.077992 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.078985 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.080497 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.081224 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.082078 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.083586 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.084498 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.085703 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.086551 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.087742 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.088269 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.088796 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.089920 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.090488 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.091556 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.091958 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.092600 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.098923 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.099521 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.100670 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.101295 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.102434 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.103452 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.104127 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.105179 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.105725 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.110995 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.111770 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.112411 4573 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.112519 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.118989 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.119590 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.120734 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.122476 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.123310 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.128030 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.128825 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.140314 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.141395 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.142972 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.143892 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.147544 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.148340 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.200173 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.205864 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-netd\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.205899 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fssxg\" (UniqueName: \"kubernetes.io/projected/48186949-d112-4fb7-aa60-026d0176c73e-kube-api-access-fssxg\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.205926 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-ovn-kubernetes\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.205947 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-systemd-units\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.205982 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-kubelet\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206004 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-ovn\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206024 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48186949-d112-4fb7-aa60-026d0176c73e-ovn-node-metrics-cert\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206042 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56192d0e-1b05-4a3d-b176-b3d27de36527-host\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206080 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-var-lib-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206098 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206123 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-env-overrides\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206144 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-node-log\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206177 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-etc-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206197 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-slash\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206216 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-log-socket\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206247 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206269 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-netns\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206285 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56192d0e-1b05-4a3d-b176-b3d27de36527-serviceca\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206304 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-config\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206324 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsm9q\" (UniqueName: \"kubernetes.io/projected/56192d0e-1b05-4a3d-b176-b3d27de36527-kube-api-access-vsm9q\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206354 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-systemd\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206373 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-bin\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206391 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-script-lib\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206648 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-netd\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206954 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-ovn-kubernetes\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.206983 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-systemd-units\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.207008 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-kubelet\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.207030 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-ovn\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.209271 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.209372 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-log-socket\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.209418 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/56192d0e-1b05-4a3d-b176-b3d27de36527-host\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.209450 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-var-lib-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.209477 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.210011 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-env-overrides\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.210245 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.210461 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-slash\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.230311 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-node-log\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.230378 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.230398 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-netns\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.231412 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/56192d0e-1b05-4a3d-b176-b3d27de36527-serviceca\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.232074 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.232646 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.233314 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-systemd\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.233422 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-bin\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.234169 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-script-lib\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.234482 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.234659 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-config\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.234982 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-etc-openvswitch\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.235502 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.236459 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.237667 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48186949-d112-4fb7-aa60-026d0176c73e-ovn-node-metrics-cert\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.241587 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.242310 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.274264 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.287870 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fssxg\" (UniqueName: \"kubernetes.io/projected/48186949-d112-4fb7-aa60-026d0176c73e-kube-api-access-fssxg\") pod \"ovnkube-node-x466z\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.298868 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05" exitCode=255 Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.298977 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.299681 4573 scope.go:117] "RemoveContainer" containerID="c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.308351 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.316312 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerStarted","Data":"cbceac120f35446082c02c5a52c760e14c71f8bd806460a5cf7c21f38a109cfd"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.328492 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsm9q\" (UniqueName: \"kubernetes.io/projected/56192d0e-1b05-4a3d-b176-b3d27de36527-kube-api-access-vsm9q\") pod \"node-ca-blvnm\" (UID: \"56192d0e-1b05-4a3d-b176-b3d27de36527\") " pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.331419 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zxsgx" event={"ID":"b73da931-0de5-4fd1-b8ff-878cc7008bf3","Type":"ContainerStarted","Data":"c429e44153fe31ae8138ac30a97315dc4336eaf9a214bf4d973819a008c0c61a"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.343503 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.343577 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.343595 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ed9b2c6da205cea60a770ba35eea76c22e3a22b50bd778349c735cc4a0c90645"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.366203 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.366271 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"b0071140f676bee8bd27ec651c8cf05d92f417593047e1b6130b6a572ddf0f51"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.384789 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ch24s" event={"ID":"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d","Type":"ContainerStarted","Data":"8d3af16c88a1351453474136cee5c224f4ae3f3f127285a846df0cc464e30520"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.400663 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.429607 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.429734 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1d909c66a1be38c54b1590de87f1746332fc8bd1272c554316793e91a649fb05"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.433298 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a10c8de206fa717aaa099e6dc7dd448753e65083b6ee2d07fb7f412e2642cbfc"} Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.502866 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.529724 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.544855 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-blvnm" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.556259 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:20 crc kubenswrapper[4573]: W1203 08:38:20.664192 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48186949_d112_4fb7_aa60_026d0176c73e.slice/crio-840a3cbc3222906512a6e6cfd20c7c4c9e8d20852872eaeab26156a645e04825 WatchSource:0}: Error finding container 840a3cbc3222906512a6e6cfd20c7c4c9e8d20852872eaeab26156a645e04825: Status 404 returned error can't find the container with id 840a3cbc3222906512a6e6cfd20c7c4c9e8d20852872eaeab26156a645e04825 Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.667353 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.713686 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.718769 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.719181 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:38:22.719150783 +0000 UTC m=+23.287530042 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.719468 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.719722 4573 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.719904 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:22.719890013 +0000 UTC m=+23.288269272 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.750740 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.797501 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.820602 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.820662 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.820699 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.820865 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.820884 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.820897 4573 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.820957 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:22.820938879 +0000 UTC m=+23.389318138 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.820990 4573 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.821018 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.821031 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.821038 4573 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.821080 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:22.821073632 +0000 UTC m=+23.389452891 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:20 crc kubenswrapper[4573]: E1203 08:38:20.821114 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:22.821092213 +0000 UTC m=+23.389471472 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:20 crc kubenswrapper[4573]: I1203 08:38:20.978123 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.040398 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:21 crc kubenswrapper[4573]: E1203 08:38:21.042469 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.040822 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:21 crc kubenswrapper[4573]: E1203 08:38:21.042782 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.155566 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.222323 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.254582 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.316854 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.359617 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.403128 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.434808 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.440777 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-blvnm" event={"ID":"56192d0e-1b05-4a3d-b176-b3d27de36527","Type":"ContainerStarted","Data":"3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.441462 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-blvnm" event={"ID":"56192d0e-1b05-4a3d-b176-b3d27de36527","Type":"ContainerStarted","Data":"91b3c16e4dbc3e0b752bbaab092d8fa40b89a7ed188c37baeca3c09ab96a0fc9"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.442905 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerStarted","Data":"821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.445785 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ch24s" event={"ID":"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d","Type":"ContainerStarted","Data":"296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.447116 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade" exitCode=0 Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.447228 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.447300 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"840a3cbc3222906512a6e6cfd20c7c4c9e8d20852872eaeab26156a645e04825"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.450151 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.452433 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.453241 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.467821 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.467893 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.469812 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zxsgx" event={"ID":"b73da931-0de5-4fd1-b8ff-878cc7008bf3","Type":"ContainerStarted","Data":"937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee"} Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.485778 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.516646 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.548651 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.577474 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.652522 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.681136 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.703846 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.721549 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.737823 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.761985 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.783207 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.802023 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.823989 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.841243 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.860158 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.879214 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.906580 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.923339 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:21 crc kubenswrapper[4573]: I1203 08:38:21.946697 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:21.991351 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.030156 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.030368 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.033801 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.065426 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.083784 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.102862 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.119456 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.144170 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.495846 4573 generic.go:334] "Generic (PLEG): container finished" podID="60e19539-4df1-4909-b8ee-c6ac0fcfc52d" containerID="821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b" exitCode=0 Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.495928 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerDied","Data":"821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b"} Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.511926 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.511973 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.511984 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.511995 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.512004 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.536912 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.569317 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.673580 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.694199 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.728313 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.766428 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.780326 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.780591 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:38:26.780549151 +0000 UTC m=+27.348928410 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.780687 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.780919 4573 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.780964 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:26.780957711 +0000 UTC m=+27.349336970 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.801959 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.824468 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.856933 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.881273 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.881328 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.881366 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.881471 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.881488 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.881499 4573 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.881540 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:26.881527974 +0000 UTC m=+27.449907233 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.881874 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.881892 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.881900 4573 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.881949 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:26.881941585 +0000 UTC m=+27.450320844 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.882259 4573 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:22 crc kubenswrapper[4573]: E1203 08:38:22.882429 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:26.882393497 +0000 UTC m=+27.450772756 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.900629 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.935399 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:22 crc kubenswrapper[4573]: I1203 08:38:22.951981 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:22Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.026892 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.029967 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:23 crc kubenswrapper[4573]: E1203 08:38:23.030211 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.030528 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:23 crc kubenswrapper[4573]: E1203 08:38:23.030643 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.519543 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerStarted","Data":"e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d"} Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.525499 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.540819 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.568363 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.587008 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.602200 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.624870 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.639133 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.646096 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.651012 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.663306 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.668643 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.686579 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.704976 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.740969 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.753616 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.811105 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.852435 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.915821 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:23 crc kubenswrapper[4573]: I1203 08:38:23.970761 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:23Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.019870 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.058346 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:24 crc kubenswrapper[4573]: E1203 08:38:24.058606 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.063627 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.104888 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.136263 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.151860 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.177876 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.195822 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.212659 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.231821 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.243359 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.255883 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.267064 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.529451 4573 generic.go:334] "Generic (PLEG): container finished" podID="60e19539-4df1-4909-b8ee-c6ac0fcfc52d" containerID="e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d" exitCode=0 Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.529681 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerDied","Data":"e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d"} Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.542702 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.561310 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.577559 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.591367 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.604985 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.617975 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.628302 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.648603 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.663359 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.681443 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.696388 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.710407 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.727321 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:24 crc kubenswrapper[4573]: I1203 08:38:24.743174 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:24Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.029692 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.029710 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:25 crc kubenswrapper[4573]: E1203 08:38:25.029839 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:25 crc kubenswrapper[4573]: E1203 08:38:25.029943 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.325369 4573 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.327134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.327172 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.327183 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.327235 4573 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.334304 4573 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.334511 4573 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.335401 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.335432 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.335443 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.335459 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.335472 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: E1203 08:38:25.350397 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.354360 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.354404 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.354413 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.354425 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.354434 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: E1203 08:38:25.364989 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.368518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.368541 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.368549 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.368563 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.368572 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: E1203 08:38:25.380190 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.383599 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.383632 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.383642 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.383657 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.383668 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: E1203 08:38:25.396296 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.399848 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.399893 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.399904 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.399919 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.399928 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: E1203 08:38:25.418285 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: E1203 08:38:25.418417 4573 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.420007 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.420092 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.420112 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.420134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.420150 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.526426 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.526499 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.526516 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.526535 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.526551 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.539165 4573 generic.go:334] "Generic (PLEG): container finished" podID="60e19539-4df1-4909-b8ee-c6ac0fcfc52d" containerID="b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733" exitCode=0 Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.539236 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerDied","Data":"b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.541140 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.546231 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.558541 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.581077 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.598215 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.612703 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.626896 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.628452 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.628507 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.628519 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.628545 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.628560 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.639972 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.654245 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.672705 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.691960 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.703758 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.716765 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.729403 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.730897 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.730939 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.730950 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.730968 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.730979 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.757688 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.767785 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.780815 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.801424 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.816985 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.833598 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.833638 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.833647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.833661 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.833671 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.834406 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.852749 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.869133 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.881754 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.895042 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.907903 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.919773 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.930787 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.935753 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.935789 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.935800 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.935815 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.935828 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:25Z","lastTransitionTime":"2025-12-03T08:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.945692 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.956459 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:25 crc kubenswrapper[4573]: I1203 08:38:25.967225 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:25Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.029508 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.029636 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.037901 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.037937 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.037956 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.037971 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.037982 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.140917 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.140953 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.140962 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.140976 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.140985 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.243556 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.243601 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.243614 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.243631 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.243644 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.345979 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.346007 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.346015 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.346026 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.346034 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.449302 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.449693 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.449907 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.450098 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.450239 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.552825 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.552895 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.552922 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.552950 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.552972 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.555789 4573 generic.go:334] "Generic (PLEG): container finished" podID="60e19539-4df1-4909-b8ee-c6ac0fcfc52d" containerID="914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43" exitCode=0 Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.555867 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerDied","Data":"914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.578787 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.605652 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.625090 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.644475 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.655700 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.656183 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.656208 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.656242 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.656261 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.664586 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.681859 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.698171 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.716747 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.733883 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.747572 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.764576 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.764849 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.764660 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.764937 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.765238 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.765260 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.780118 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.802801 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.815093 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:26Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.837272 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.837446 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.837483 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:38:34.83745245 +0000 UTC m=+35.405831719 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.837658 4573 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.837806 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:34.837779389 +0000 UTC m=+35.406158688 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.868281 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.868322 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.868337 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.868357 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.868370 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.938078 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.938266 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.938387 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.938286 4573 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.938647 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:34.938627559 +0000 UTC m=+35.507006818 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.938491 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.938809 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.938890 4573 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.938983 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:34.938971869 +0000 UTC m=+35.507351128 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.938400 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.939207 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.939280 4573 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:26 crc kubenswrapper[4573]: E1203 08:38:26.939380 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:34.939369899 +0000 UTC m=+35.507749158 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.971684 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.971749 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.971760 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.972560 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:26 crc kubenswrapper[4573]: I1203 08:38:26.972612 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:26Z","lastTransitionTime":"2025-12-03T08:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.029830 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.029940 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:27 crc kubenswrapper[4573]: E1203 08:38:27.030098 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:27 crc kubenswrapper[4573]: E1203 08:38:27.030334 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.076101 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.076126 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.076141 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.076157 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.076166 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.178421 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.178469 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.178482 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.178502 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.178518 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.280775 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.280822 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.280836 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.280854 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.280866 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.384243 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.384295 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.384309 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.384330 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.384345 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.487151 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.487201 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.487214 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.487231 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.487245 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.563454 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.563774 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.567444 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerStarted","Data":"4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.575387 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.597142 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.597913 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.597948 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.597960 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.597975 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.597985 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.621684 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.640089 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.651296 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.666735 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.671913 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.682520 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.696511 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.700665 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.700702 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.700716 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.700734 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.700778 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.711939 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.729982 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.741429 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.753003 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.765891 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.776323 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.790968 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.803237 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.803271 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.803283 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.803300 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.803311 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.811471 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.824694 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.838701 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.851164 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.863170 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.874644 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.884723 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.896332 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.905531 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.905579 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.905597 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.905621 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.905642 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:27Z","lastTransitionTime":"2025-12-03T08:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.907076 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.916170 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.928505 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.939619 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:27 crc kubenswrapper[4573]: I1203 08:38:27.963387 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.012985 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.013026 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.013035 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.013081 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.013123 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.029453 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:28 crc kubenswrapper[4573]: E1203 08:38:28.029597 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.115743 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.115782 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.115793 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.115808 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.115819 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.218488 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.218527 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.218538 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.218553 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.218563 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.321091 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.321137 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.321146 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.321162 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.321173 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.423809 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.423871 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.423887 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.423908 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.423921 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.526085 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.526147 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.526156 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.526170 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.526180 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.576556 4573 generic.go:334] "Generic (PLEG): container finished" podID="60e19539-4df1-4909-b8ee-c6ac0fcfc52d" containerID="4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93" exitCode=0 Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.576625 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerDied","Data":"4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.576713 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.577657 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.599460 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.601343 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.612778 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.627528 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.628180 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.628209 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.628217 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.628231 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.628239 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.641526 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.653166 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.663953 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.673604 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.699968 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.713653 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.728312 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.731506 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.731544 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.731556 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.731574 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.731599 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.743826 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.758126 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.770628 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.783110 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.793157 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.817105 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.830193 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.833564 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.833600 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.833610 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.833623 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.833632 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.843435 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.855577 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.867992 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.879558 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.890633 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.899553 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.914607 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.926551 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.937565 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.937616 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.937625 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.937645 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.937653 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:28Z","lastTransitionTime":"2025-12-03T08:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.942034 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.956408 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:28 crc kubenswrapper[4573]: I1203 08:38:28.968902 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:28Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.029401 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:29 crc kubenswrapper[4573]: E1203 08:38:29.029572 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.029800 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:29 crc kubenswrapper[4573]: E1203 08:38:29.030018 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.039204 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.039239 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.039248 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.039261 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.039271 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.141811 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.142287 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.142413 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.142540 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.142648 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.247698 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.248112 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.248220 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.248331 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.248424 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.351301 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.351347 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.351367 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.351387 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.351399 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.491775 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.491826 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.491835 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.491850 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.491860 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.582123 4573 generic.go:334] "Generic (PLEG): container finished" podID="60e19539-4df1-4909-b8ee-c6ac0fcfc52d" containerID="d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d" exitCode=0 Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.582250 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.582234 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerDied","Data":"d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.594637 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.594685 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.594697 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.594717 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.594728 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.604820 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.622927 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.635509 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.650897 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.663369 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.677285 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.688352 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.697007 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.697039 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.697070 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.697085 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.697096 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.713513 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.732076 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.749980 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.764364 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.776096 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.788575 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.799417 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.799470 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.799484 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.799504 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.799517 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.800595 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:29Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.902375 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.902412 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.902424 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.902444 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:29 crc kubenswrapper[4573]: I1203 08:38:29.902454 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:29Z","lastTransitionTime":"2025-12-03T08:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.007256 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.007304 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.007317 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.007337 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.007350 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.030608 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:30 crc kubenswrapper[4573]: E1203 08:38:30.030725 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.044868 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.056596 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.069942 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.081717 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.094177 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.109135 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.109179 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.109189 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.109205 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.109215 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.116190 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.136735 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.155378 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.169568 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.180462 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.189654 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.210972 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.211011 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.211023 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.211039 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.211089 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.224190 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.240498 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.258435 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.314313 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.314350 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.314360 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.314375 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.314384 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.416704 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.416733 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.416742 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.416756 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.416764 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.520097 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.520152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.520169 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.520188 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.520205 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.560166 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.596422 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" event={"ID":"60e19539-4df1-4909-b8ee-c6ac0fcfc52d","Type":"ContainerStarted","Data":"e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.616761 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.624368 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.624410 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.624421 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.624437 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.624447 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.631150 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.647644 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.662457 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.675037 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.686719 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.704351 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.720937 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.726754 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.726827 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.726843 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.726862 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.726878 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.736006 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.750751 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.763449 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.783592 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.797108 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.817832 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.830031 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.830101 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.830114 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.830137 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.830149 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.932327 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.932367 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.932399 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.932412 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:30 crc kubenswrapper[4573]: I1203 08:38:30.932420 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:30Z","lastTransitionTime":"2025-12-03T08:38:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.029608 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.029701 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:31 crc kubenswrapper[4573]: E1203 08:38:31.029737 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:31 crc kubenswrapper[4573]: E1203 08:38:31.029938 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.035498 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.035532 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.035545 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.035562 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.035574 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.098288 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.121242 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.138800 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.138841 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.138853 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.138869 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.138880 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.143504 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.164187 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.189703 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.207379 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.227407 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.237579 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.241030 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.241104 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.241154 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.241172 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.241184 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.266081 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.282846 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.296115 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.313614 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.330872 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.344195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.344257 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.344269 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.344295 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.344308 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.348455 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.361926 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.446895 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.446972 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.446998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.447030 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.447095 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.550215 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.550771 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.550802 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.550822 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.550833 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.598572 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/0.log" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.600698 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763" exitCode=1 Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.600816 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.601486 4573 scope.go:117] "RemoveContainer" containerID="d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.620606 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.638094 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.651540 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.652918 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.653584 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.653735 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.653834 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.654037 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.665951 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.676667 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.688247 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.699410 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.718545 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"message\\\":\\\"] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 08:38:30.586542 5714 factory.go:656] Stopping watch factory\\\\nI1203 08:38:30.586568 5714 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 08:38:30.586577 5714 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:30.586584 5714 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:30.586593 5714 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 08:38:30.586604 5714 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.586790 5714 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587108 5714 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587400 5714 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.587391 5714 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.733268 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.746361 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.756255 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.756462 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.756563 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.756650 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.756718 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.761397 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.772778 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.784869 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.796692 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:31Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.859757 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.860198 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.860326 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.860420 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.860497 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.964152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.964460 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.964547 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.964643 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:31 crc kubenswrapper[4573]: I1203 08:38:31.964744 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:31Z","lastTransitionTime":"2025-12-03T08:38:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.030330 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:32 crc kubenswrapper[4573]: E1203 08:38:32.030462 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.067603 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.067647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.067660 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.067679 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.067692 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.170292 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.170335 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.170345 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.170361 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.170375 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.273358 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.273636 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.273842 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.274140 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.274269 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.379419 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.379475 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.379489 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.379511 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.379526 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.482332 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.482368 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.482380 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.482396 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.482441 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.584727 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.584796 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.584809 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.584828 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.584840 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.627085 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/0.log" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.687203 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.687251 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.687261 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.687273 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.687281 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.789743 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.789774 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.789782 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.789795 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.789803 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.891623 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.891669 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.891682 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.891700 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.891711 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.995246 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.995296 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.995308 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.995326 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:32 crc kubenswrapper[4573]: I1203 08:38:32.995337 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:32Z","lastTransitionTime":"2025-12-03T08:38:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.030385 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.030394 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:33 crc kubenswrapper[4573]: E1203 08:38:33.030584 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:33 crc kubenswrapper[4573]: E1203 08:38:33.030643 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.098996 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.099114 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.099131 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.099152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.099166 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.202017 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.202084 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.202100 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.202119 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.202133 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.305424 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.305476 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.305509 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.305525 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.305535 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.408799 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.408838 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.408847 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.408864 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.408872 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.510401 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.510429 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.510436 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.510448 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.510457 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.613305 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.613349 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.613362 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.613394 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.613403 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.635197 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/0.log" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.638209 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.639171 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.651605 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.661390 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.673448 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.686711 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.698684 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.710527 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.715446 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.715479 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.715489 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.715504 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.715514 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.722320 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.747913 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"message\\\":\\\"] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 08:38:30.586542 5714 factory.go:656] Stopping watch factory\\\\nI1203 08:38:30.586568 5714 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 08:38:30.586577 5714 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:30.586584 5714 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:30.586593 5714 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 08:38:30.586604 5714 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.586790 5714 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587108 5714 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587400 5714 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.587391 5714 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.762748 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.789291 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.805780 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n"] Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.806198 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.809257 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.811204 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.817762 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.817795 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.817809 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.817824 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.817835 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.834624 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dx9r\" (UniqueName: \"kubernetes.io/projected/bdff5b8e-11e4-4def-a585-842ecb30f940-kube-api-access-8dx9r\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.834667 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bdff5b8e-11e4-4def-a585-842ecb30f940-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.834703 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bdff5b8e-11e4-4def-a585-842ecb30f940-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.834721 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bdff5b8e-11e4-4def-a585-842ecb30f940-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.841301 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.857964 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.874886 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.885676 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.893850 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.913798 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"message\\\":\\\"] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 08:38:30.586542 5714 factory.go:656] Stopping watch factory\\\\nI1203 08:38:30.586568 5714 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 08:38:30.586577 5714 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:30.586584 5714 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:30.586593 5714 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 08:38:30.586604 5714 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.586790 5714 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587108 5714 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587400 5714 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.587391 5714 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.920529 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.920560 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.920569 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.920582 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.920591 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:33Z","lastTransitionTime":"2025-12-03T08:38:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.931712 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.935459 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dx9r\" (UniqueName: \"kubernetes.io/projected/bdff5b8e-11e4-4def-a585-842ecb30f940-kube-api-access-8dx9r\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.935510 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bdff5b8e-11e4-4def-a585-842ecb30f940-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.935556 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bdff5b8e-11e4-4def-a585-842ecb30f940-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.935584 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bdff5b8e-11e4-4def-a585-842ecb30f940-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.936311 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bdff5b8e-11e4-4def-a585-842ecb30f940-env-overrides\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.936399 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bdff5b8e-11e4-4def-a585-842ecb30f940-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.942307 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bdff5b8e-11e4-4def-a585-842ecb30f940-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.951081 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.953135 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dx9r\" (UniqueName: \"kubernetes.io/projected/bdff5b8e-11e4-4def-a585-842ecb30f940-kube-api-access-8dx9r\") pod \"ovnkube-control-plane-749d76644c-q298n\" (UID: \"bdff5b8e-11e4-4def-a585-842ecb30f940\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.963256 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.978671 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:33 crc kubenswrapper[4573]: I1203 08:38:33.991868 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:33Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.004666 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.014265 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.022776 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.022821 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.022849 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.022866 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.022876 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.026259 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.029473 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.029563 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.043254 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.056023 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.068399 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.081703 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.092287 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.121696 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.129094 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.129121 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.129129 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.129142 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.129150 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: W1203 08:38:34.139904 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdff5b8e_11e4_4def_a585_842ecb30f940.slice/crio-d2d221a4b4d2bb4fb583b5e69554df295e9e9a2ced76d40bdf81796b027ba132 WatchSource:0}: Error finding container d2d221a4b4d2bb4fb583b5e69554df295e9e9a2ced76d40bdf81796b027ba132: Status 404 returned error can't find the container with id d2d221a4b4d2bb4fb583b5e69554df295e9e9a2ced76d40bdf81796b027ba132 Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.231693 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.231721 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.231729 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.231741 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.231750 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.333736 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.333773 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.333782 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.333820 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.333830 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.439946 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.439975 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.440004 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.440017 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.440026 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.542114 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.542165 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.542176 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.542195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.542206 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.570664 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kcphv"] Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.571307 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.571365 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.589492 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.602548 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.619657 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.637153 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.641778 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.641844 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vchs5\" (UniqueName: \"kubernetes.io/projected/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-kube-api-access-vchs5\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.643921 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" event={"ID":"bdff5b8e-11e4-4def-a585-842ecb30f940","Type":"ContainerStarted","Data":"f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.644028 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" event={"ID":"bdff5b8e-11e4-4def-a585-842ecb30f940","Type":"ContainerStarted","Data":"58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.644120 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" event={"ID":"bdff5b8e-11e4-4def-a585-842ecb30f940","Type":"ContainerStarted","Data":"d2d221a4b4d2bb4fb583b5e69554df295e9e9a2ced76d40bdf81796b027ba132"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.645146 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.645172 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.645181 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.645195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.645204 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.646132 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/1.log" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.646777 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/0.log" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.649078 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc" exitCode=1 Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.649132 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.649175 4573 scope.go:117] "RemoveContainer" containerID="d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.649998 4573 scope.go:117] "RemoveContainer" containerID="01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc" Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.650215 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.656103 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.673693 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.685839 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.697767 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.712256 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.726503 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.741999 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.742565 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vchs5\" (UniqueName: \"kubernetes.io/projected/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-kube-api-access-vchs5\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.742697 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.742925 4573 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.743014 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs podName:bb7f6603-0f41-4e0f-b77e-47dbcc676aaf nodeName:}" failed. No retries permitted until 2025-12-03 08:38:35.242989133 +0000 UTC m=+35.811368382 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs") pod "network-metrics-daemon-kcphv" (UID: "bb7f6603-0f41-4e0f-b77e-47dbcc676aaf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.747521 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.747561 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.747570 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.747588 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.747601 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.757403 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.769524 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vchs5\" (UniqueName: \"kubernetes.io/projected/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-kube-api-access-vchs5\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.769756 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.789639 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"message\\\":\\\"] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 08:38:30.586542 5714 factory.go:656] Stopping watch factory\\\\nI1203 08:38:30.586568 5714 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 08:38:30.586577 5714 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:30.586584 5714 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:30.586593 5714 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 08:38:30.586604 5714 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.586790 5714 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587108 5714 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587400 5714 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.587391 5714 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.805955 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.821116 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.843977 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.844247 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:38:50.844208953 +0000 UTC m=+51.412588222 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.844527 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.844789 4573 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.844907 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:50.844887651 +0000 UTC m=+51.413266910 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.849036 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d06f581f985d577a51006f24860f0f06231df144158a7083c330eb1f30d03763\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"message\\\":\\\"] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 08:38:30.586542 5714 factory.go:656] Stopping watch factory\\\\nI1203 08:38:30.586568 5714 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 08:38:30.586577 5714 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:30.586584 5714 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:30.586593 5714 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 08:38:30.586604 5714 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.586790 5714 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587108 5714 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:30.587400 5714 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:30.587391 5714 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.967375 5907 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:33.967566 5907 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.968341 5907 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 08:38:33.968367 5907 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 08:38:33.968394 5907 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 08:38:33.968399 5907 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 08:38:33.968403 5907 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:33.968412 5907 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 08:38:33.968417 5907 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:33.968425 5907 factory.go:656] Stopping watch factory\\\\nI1203 08:38:33.968426 5907 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 08:38:33.968436 5907 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 08:38:33.968436 5907 ovnkube.go:599] Stopped ovnkube\\\\nI1203 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.850254 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.850370 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.850484 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.850616 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.850730 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.861107 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.876081 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.891876 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.906122 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.924899 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.941162 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.945271 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.945315 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.945347 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945465 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945483 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945493 4573 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945536 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:50.945521295 +0000 UTC m=+51.513900554 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945550 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945611 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945631 4573 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945697 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:50.945677389 +0000 UTC m=+51.514056658 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945757 4573 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: E1203 08:38:34.945787 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:38:50.945778452 +0000 UTC m=+51.514157721 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.953676 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.953700 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.953708 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.953721 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.953729 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:34Z","lastTransitionTime":"2025-12-03T08:38:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.957295 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.971541 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:34 crc kubenswrapper[4573]: I1203 08:38:34.986824 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.000240 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:34Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.012817 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.023763 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.029483 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.029529 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.029627 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.029727 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.035346 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.051318 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.056558 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.056614 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.056636 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.056666 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.056684 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.067809 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.159068 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.159121 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.159134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.159153 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.159165 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.248748 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.248922 4573 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.248975 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs podName:bb7f6603-0f41-4e0f-b77e-47dbcc676aaf nodeName:}" failed. No retries permitted until 2025-12-03 08:38:36.248959531 +0000 UTC m=+36.817338790 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs") pod "network-metrics-daemon-kcphv" (UID: "bb7f6603-0f41-4e0f-b77e-47dbcc676aaf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.261481 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.261528 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.261540 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.261560 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.261574 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.365579 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.365638 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.365653 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.365675 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.365690 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.468800 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.468856 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.468870 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.468893 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.468906 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.571906 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.571952 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.571967 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.571987 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.571999 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.655036 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/1.log" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.660163 4573 scope.go:117] "RemoveContainer" containerID="01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.660465 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.674291 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.674339 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.674353 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.674371 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.674387 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.678559 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.693075 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.712096 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.730558 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.743318 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.762391 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.776195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.776216 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.776225 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.776237 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.776245 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.777629 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.789039 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.806463 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.806536 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.806553 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.806570 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.806583 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.811624 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.967375 5907 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:33.967566 5907 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.968341 5907 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 08:38:33.968367 5907 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 08:38:33.968394 5907 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 08:38:33.968399 5907 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 08:38:33.968403 5907 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:33.968412 5907 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 08:38:33.968417 5907 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:33.968425 5907 factory.go:656] Stopping watch factory\\\\nI1203 08:38:33.968426 5907 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 08:38:33.968436 5907 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 08:38:33.968436 5907 ovnkube.go:599] Stopped ovnkube\\\\nI1203 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.824155 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.832207 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.832264 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.832275 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.832292 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.832303 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.837415 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.844665 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.848916 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.848974 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.848989 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.849013 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.849029 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.852513 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.864691 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.869285 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.869454 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.869540 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.869671 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.869819 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.869568 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.884565 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.886171 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.888988 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.889022 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.889038 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.889090 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.889104 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.901222 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: E1203 08:38:35.901685 4573 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.903293 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.904098 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.904202 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.904290 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.904377 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.904459 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:35Z","lastTransitionTime":"2025-12-03T08:38:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.918792 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:35 crc kubenswrapper[4573]: I1203 08:38:35.932593 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:35Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.007488 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.007656 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.007850 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.007996 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.008161 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.029942 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.030030 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:36 crc kubenswrapper[4573]: E1203 08:38:36.030426 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:36 crc kubenswrapper[4573]: E1203 08:38:36.030588 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.111238 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.111265 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.111276 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.111289 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.111297 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.214976 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.215328 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.215402 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.215475 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.215534 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.259114 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:36 crc kubenswrapper[4573]: E1203 08:38:36.259405 4573 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:36 crc kubenswrapper[4573]: E1203 08:38:36.259597 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs podName:bb7f6603-0f41-4e0f-b77e-47dbcc676aaf nodeName:}" failed. No retries permitted until 2025-12-03 08:38:38.259560379 +0000 UTC m=+38.827939678 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs") pod "network-metrics-daemon-kcphv" (UID: "bb7f6603-0f41-4e0f-b77e-47dbcc676aaf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.317931 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.317988 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.318001 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.318022 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.318035 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.422361 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.422435 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.422461 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.422491 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.422513 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.525627 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.525703 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.525731 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.525765 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.525789 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.629329 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.629372 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.629385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.629404 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.629416 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.732385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.732426 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.732439 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.732459 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.732472 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.835143 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.835186 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.835197 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.835215 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.835230 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.937838 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.937883 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.937893 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.937909 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:36 crc kubenswrapper[4573]: I1203 08:38:36.937918 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:36Z","lastTransitionTime":"2025-12-03T08:38:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.030184 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.030252 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:37 crc kubenswrapper[4573]: E1203 08:38:37.030363 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:37 crc kubenswrapper[4573]: E1203 08:38:37.030494 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.041363 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.041439 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.041452 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.041475 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.041489 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.143889 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.143962 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.143974 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.143992 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.144005 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.251666 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.251798 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.251815 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.251836 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.251853 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.355536 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.355590 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.355600 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.355617 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.355628 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.458271 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.458312 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.458325 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.458345 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.458356 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.561712 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.561775 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.561786 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.561814 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.561830 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.667015 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.667556 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.667813 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.667981 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.668203 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.771284 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.771349 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.771368 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.771425 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.771445 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.874993 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.875095 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.875123 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.875156 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.875181 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.978354 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.978584 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.978709 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.978871 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:37 crc kubenswrapper[4573]: I1203 08:38:37.979008 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:37Z","lastTransitionTime":"2025-12-03T08:38:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.030193 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.030200 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:38 crc kubenswrapper[4573]: E1203 08:38:38.030986 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:38 crc kubenswrapper[4573]: E1203 08:38:38.031195 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.082575 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.082870 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.083130 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.083326 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.083528 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.186883 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.187297 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.187494 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.187701 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.187948 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.280352 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:38 crc kubenswrapper[4573]: E1203 08:38:38.280529 4573 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:38 crc kubenswrapper[4573]: E1203 08:38:38.280599 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs podName:bb7f6603-0f41-4e0f-b77e-47dbcc676aaf nodeName:}" failed. No retries permitted until 2025-12-03 08:38:42.280580424 +0000 UTC m=+42.848959693 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs") pod "network-metrics-daemon-kcphv" (UID: "bb7f6603-0f41-4e0f-b77e-47dbcc676aaf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.291385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.291636 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.291814 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.292042 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.292303 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.395674 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.396024 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.396284 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.396495 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.396631 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.499928 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.499969 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.499978 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.499996 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.500026 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.602828 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.603178 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.603270 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.603366 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.603447 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.706080 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.706426 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.706611 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.706740 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.706864 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.810175 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.811003 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.811125 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.811265 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.811361 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.914328 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.914366 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.914376 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.914397 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:38 crc kubenswrapper[4573]: I1203 08:38:38.914409 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:38Z","lastTransitionTime":"2025-12-03T08:38:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.017399 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.017461 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.017481 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.017502 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.017518 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.029753 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:39 crc kubenswrapper[4573]: E1203 08:38:39.029912 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.030181 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:39 crc kubenswrapper[4573]: E1203 08:38:39.030309 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.121017 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.121099 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.121114 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.121135 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.121147 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.223920 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.223992 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.224005 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.224018 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.224026 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.326737 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.326798 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.326810 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.326837 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.326848 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.430340 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.430405 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.430420 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.430466 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.430482 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.533546 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.533622 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.533646 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.533690 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.533713 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.636900 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.636949 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.636998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.637021 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.637036 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.739720 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.739769 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.739783 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.739802 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.739815 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.843534 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.843592 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.843609 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.843629 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.843640 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.946657 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.946702 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.946719 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.946740 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:39 crc kubenswrapper[4573]: I1203 08:38:39.946757 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:39Z","lastTransitionTime":"2025-12-03T08:38:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.030353 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.030477 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:40 crc kubenswrapper[4573]: E1203 08:38:40.030686 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:40 crc kubenswrapper[4573]: E1203 08:38:40.031166 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.050023 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.050115 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.050134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.050156 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.050175 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.059607 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.078604 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.096841 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.118476 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.132508 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.148527 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.152612 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.152650 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.152663 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.152680 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.152691 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.161400 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.174459 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.187859 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.205098 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.218556 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.231588 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.246820 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.255427 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.255647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.255755 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.255885 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.256014 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.262265 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.274113 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.297691 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.967375 5907 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:33.967566 5907 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.968341 5907 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 08:38:33.968367 5907 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 08:38:33.968394 5907 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 08:38:33.968399 5907 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 08:38:33.968403 5907 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:33.968412 5907 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 08:38:33.968417 5907 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:33.968425 5907 factory.go:656] Stopping watch factory\\\\nI1203 08:38:33.968426 5907 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 08:38:33.968436 5907 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 08:38:33.968436 5907 ovnkube.go:599] Stopped ovnkube\\\\nI1203 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:40Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.358832 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.358886 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.358902 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.358923 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.358937 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.468228 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.468271 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.468450 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.468475 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.468490 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.570896 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.570939 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.570950 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.570966 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.570977 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.673648 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.673680 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.673690 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.673703 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.673713 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.776750 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.776805 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.776819 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.776839 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.776851 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.880180 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.880256 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.880274 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.880299 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.880316 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.984502 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.984573 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.984591 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.984617 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:40 crc kubenswrapper[4573]: I1203 08:38:40.984638 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:40Z","lastTransitionTime":"2025-12-03T08:38:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.029684 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.029808 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:41 crc kubenswrapper[4573]: E1203 08:38:41.029885 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:41 crc kubenswrapper[4573]: E1203 08:38:41.029999 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.088131 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.088190 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.088208 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.088231 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.088246 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.191671 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.191721 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.191733 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.191753 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.191765 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.294772 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.294806 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.294814 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.294826 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.294834 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.397926 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.398088 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.398123 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.398142 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.398154 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.501175 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.501236 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.501253 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.501277 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.501293 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.603792 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.603840 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.603852 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.603871 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.603883 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.708372 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.708434 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.708452 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.708481 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.708498 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.811781 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.811835 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.811853 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.811879 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.811898 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.914569 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.914635 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.914653 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.914685 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:41 crc kubenswrapper[4573]: I1203 08:38:41.914701 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:41Z","lastTransitionTime":"2025-12-03T08:38:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.017615 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.017693 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.017719 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.017749 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.017776 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.030263 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.030297 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:42 crc kubenswrapper[4573]: E1203 08:38:42.030631 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:42 crc kubenswrapper[4573]: E1203 08:38:42.030564 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.120502 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.120575 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.120603 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.120633 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.120654 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.224123 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.224192 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.224217 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.224247 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.224269 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.327494 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.327555 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.327566 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.327586 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.327605 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.343756 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:42 crc kubenswrapper[4573]: E1203 08:38:42.343998 4573 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:42 crc kubenswrapper[4573]: E1203 08:38:42.344151 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs podName:bb7f6603-0f41-4e0f-b77e-47dbcc676aaf nodeName:}" failed. No retries permitted until 2025-12-03 08:38:50.344124784 +0000 UTC m=+50.912504043 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs") pod "network-metrics-daemon-kcphv" (UID: "bb7f6603-0f41-4e0f-b77e-47dbcc676aaf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.430833 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.430907 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.430918 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.430944 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.430957 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.534001 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.534115 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.534142 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.534171 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.534192 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.637274 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.637318 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.637328 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.637346 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.637356 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.740424 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.740496 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.740515 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.740540 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.740558 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.845896 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.845988 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.846018 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.846093 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.846134 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.949755 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.949836 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.949861 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.949895 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:42 crc kubenswrapper[4573]: I1203 08:38:42.949921 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:42Z","lastTransitionTime":"2025-12-03T08:38:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.029498 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.029531 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:43 crc kubenswrapper[4573]: E1203 08:38:43.029816 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:43 crc kubenswrapper[4573]: E1203 08:38:43.029681 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.053149 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.053215 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.053240 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.053269 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.053292 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.155741 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.155786 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.155823 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.155839 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.155851 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.259299 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.259359 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.259372 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.259391 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.259406 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.361964 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.362012 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.362025 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.362041 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.362080 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.465224 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.465297 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.465312 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.465338 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.465357 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.568723 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.568768 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.568776 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.568790 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.568801 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.672295 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.672343 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.672353 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.672369 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.672379 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.776019 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.776110 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.776128 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.776152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.776169 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.878954 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.879036 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.879069 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.879088 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.879098 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.982009 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.982066 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.982076 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.982091 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:43 crc kubenswrapper[4573]: I1203 08:38:43.982101 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:43Z","lastTransitionTime":"2025-12-03T08:38:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.030432 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.030561 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:44 crc kubenswrapper[4573]: E1203 08:38:44.030677 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:44 crc kubenswrapper[4573]: E1203 08:38:44.030746 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.085424 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.085492 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.085505 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.085527 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.085546 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.188135 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.188186 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.188203 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.188225 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.188245 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.291144 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.291183 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.291196 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.291211 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.291220 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.393505 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.393557 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.393569 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.393585 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.393597 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.496027 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.496112 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.496124 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.496142 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.496153 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.599205 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.599257 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.599270 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.599286 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.599296 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.701660 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.701731 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.701744 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.701761 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.701773 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.805138 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.805193 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.805209 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.805233 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.805251 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.907442 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.907483 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.907501 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.907516 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:44 crc kubenswrapper[4573]: I1203 08:38:44.907525 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:44Z","lastTransitionTime":"2025-12-03T08:38:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.009465 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.009503 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.009515 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.009554 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.009569 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.029712 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.029779 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:45 crc kubenswrapper[4573]: E1203 08:38:45.029856 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:45 crc kubenswrapper[4573]: E1203 08:38:45.029973 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.112425 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.112510 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.112536 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.112567 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.112591 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.215298 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.215350 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.215369 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.215389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.215402 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.317554 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.317605 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.317621 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.317643 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.317659 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.420192 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.420232 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.420244 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.420261 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.420275 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.523228 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.523303 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.523326 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.523354 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.523388 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.626463 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.626512 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.626539 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.626568 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.626580 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.729582 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.729620 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.729631 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.729646 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.729658 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.832239 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.832309 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.832323 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.832348 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.832362 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.934547 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.934584 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.934595 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.934610 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:45 crc kubenswrapper[4573]: I1203 08:38:45.934619 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:45Z","lastTransitionTime":"2025-12-03T08:38:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.029862 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:46 crc kubenswrapper[4573]: E1203 08:38:46.030057 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.029869 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:46 crc kubenswrapper[4573]: E1203 08:38:46.030138 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.039330 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.039367 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.039376 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.039390 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.039400 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.142401 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.142446 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.142457 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.142472 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.142505 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.227422 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.227458 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.227467 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.227485 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.227496 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: E1203 08:38:46.247569 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.253198 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.253407 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.253419 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.253434 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.253443 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: E1203 08:38:46.265076 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.269277 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.269318 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.269330 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.269346 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.269356 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: E1203 08:38:46.283361 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.287877 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.287930 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.287954 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.287978 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.287993 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: E1203 08:38:46.302884 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.307716 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.307758 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.307767 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.307780 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.307789 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: E1203 08:38:46.321202 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:46Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:46 crc kubenswrapper[4573]: E1203 08:38:46.321450 4573 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.323270 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.323330 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.323341 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.323354 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.323363 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.425614 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.425644 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.425653 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.425666 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.425675 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.528181 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.528223 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.528233 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.528248 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.528258 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.631577 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.631618 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.631635 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.631650 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.631659 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.735031 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.735103 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.735119 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.735139 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.735151 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.838010 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.838127 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.838139 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.838152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.838162 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.941243 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.941322 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.941357 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.941387 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:46 crc kubenswrapper[4573]: I1203 08:38:46.941409 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:46Z","lastTransitionTime":"2025-12-03T08:38:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.030086 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.030149 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:47 crc kubenswrapper[4573]: E1203 08:38:47.030327 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:47 crc kubenswrapper[4573]: E1203 08:38:47.030527 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.044332 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.044381 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.044395 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.044414 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.044429 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.147331 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.147406 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.147418 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.147439 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.147454 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.250385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.250462 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.250482 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.250506 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.250524 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.353828 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.353870 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.353882 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.353900 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.353915 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.456895 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.456937 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.456949 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.456967 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.456979 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.559864 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.559902 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.559911 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.559926 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.559936 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.662398 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.662464 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.662478 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.662516 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.662527 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.765304 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.765354 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.765365 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.765382 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.765393 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.868395 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.868452 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.868465 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.868482 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.868867 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.972209 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.972310 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.972331 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.972358 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:47 crc kubenswrapper[4573]: I1203 08:38:47.972375 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:47Z","lastTransitionTime":"2025-12-03T08:38:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.030127 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:48 crc kubenswrapper[4573]: E1203 08:38:48.030291 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.030568 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:48 crc kubenswrapper[4573]: E1203 08:38:48.030852 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.075028 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.075090 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.075274 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.075314 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.075327 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.096659 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.104654 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.112125 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.126485 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.142210 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.154464 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.166497 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.179705 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.180070 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.180116 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.180164 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.180181 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.180192 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.191680 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.203969 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.224693 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.967375 5907 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:33.967566 5907 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.968341 5907 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 08:38:33.968367 5907 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 08:38:33.968394 5907 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 08:38:33.968399 5907 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 08:38:33.968403 5907 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:33.968412 5907 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 08:38:33.968417 5907 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:33.968425 5907 factory.go:656] Stopping watch factory\\\\nI1203 08:38:33.968426 5907 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 08:38:33.968436 5907 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 08:38:33.968436 5907 ovnkube.go:599] Stopped ovnkube\\\\nI1203 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.237197 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.250334 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.262807 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.276648 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.282943 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.282993 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.283011 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.283034 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.283064 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.289992 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.305695 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.322552 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:48Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.385775 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.385849 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.385863 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.385883 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.385895 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.488568 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.488622 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.488631 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.488647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.488657 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.590708 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.590746 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.590755 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.590771 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.590780 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.692844 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.692902 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.692919 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.692942 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.692959 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.795618 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.795689 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.795715 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.795742 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.795764 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.899411 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.899462 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.899474 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.899493 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:48 crc kubenswrapper[4573]: I1203 08:38:48.899516 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:48Z","lastTransitionTime":"2025-12-03T08:38:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.001441 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.001481 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.001490 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.001504 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.001513 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.029395 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.029460 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:49 crc kubenswrapper[4573]: E1203 08:38:49.029542 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:49 crc kubenswrapper[4573]: E1203 08:38:49.029687 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.103734 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.103796 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.103807 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.103819 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.103828 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.206123 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.206158 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.206167 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.206183 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.206195 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.309338 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.309382 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.309393 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.309410 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.309420 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.413299 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.413353 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.413370 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.413391 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.413407 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.516741 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.516822 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.516840 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.516862 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.516875 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.620109 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.620173 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.620186 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.620212 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.620228 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.723267 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.723331 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.723346 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.723389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.723403 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.826539 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.826573 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.826585 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.826602 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.826615 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.931173 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.931221 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.931239 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.931262 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:49 crc kubenswrapper[4573]: I1203 08:38:49.931279 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:49Z","lastTransitionTime":"2025-12-03T08:38:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.030201 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:50 crc kubenswrapper[4573]: E1203 08:38:50.030362 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.030410 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.031184 4573 scope.go:117] "RemoveContainer" containerID="01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc" Dec 03 08:38:50 crc kubenswrapper[4573]: E1203 08:38:50.031409 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.034211 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.034247 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.034258 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.035134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.035149 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.050158 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.066694 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.083321 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.095830 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.113694 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.127336 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.139161 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.139561 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.139644 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.139708 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.139776 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.139835 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.148410 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.166575 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.967375 5907 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:33.967566 5907 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.968341 5907 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 08:38:33.968367 5907 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 08:38:33.968394 5907 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 08:38:33.968399 5907 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 08:38:33.968403 5907 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:33.968412 5907 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 08:38:33.968417 5907 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:33.968425 5907 factory.go:656] Stopping watch factory\\\\nI1203 08:38:33.968426 5907 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 08:38:33.968436 5907 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 08:38:33.968436 5907 ovnkube.go:599] Stopped ovnkube\\\\nI1203 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.180159 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.195385 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.213065 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.226963 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.242389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.242417 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.242427 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.242441 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.242452 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.245032 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.256641 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.269547 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.281211 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.344184 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.344220 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.344230 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.344243 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.344253 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.436302 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:50 crc kubenswrapper[4573]: E1203 08:38:50.436449 4573 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:50 crc kubenswrapper[4573]: E1203 08:38:50.436535 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs podName:bb7f6603-0f41-4e0f-b77e-47dbcc676aaf nodeName:}" failed. No retries permitted until 2025-12-03 08:39:06.436516324 +0000 UTC m=+67.004895583 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs") pod "network-metrics-daemon-kcphv" (UID: "bb7f6603-0f41-4e0f-b77e-47dbcc676aaf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.446091 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.446116 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.446123 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.446135 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.446143 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.548429 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.548470 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.548479 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.548492 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.548500 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.651152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.651200 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.651215 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.651234 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.651248 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.710036 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/1.log" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.711969 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.712739 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.724064 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.734623 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.749126 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.753201 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.753229 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.753239 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.753253 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.753265 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.762658 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.777799 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.788515 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.799308 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.810333 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.830167 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.967375 5907 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:33.967566 5907 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.968341 5907 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 08:38:33.968367 5907 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 08:38:33.968394 5907 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 08:38:33.968399 5907 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 08:38:33.968403 5907 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:33.968412 5907 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 08:38:33.968417 5907 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:33.968425 5907 factory.go:656] Stopping watch factory\\\\nI1203 08:38:33.968426 5907 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 08:38:33.968436 5907 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 08:38:33.968436 5907 ovnkube.go:599] Stopped ovnkube\\\\nI1203 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.842981 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.855108 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.855144 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.855153 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.855166 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.855175 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.857659 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.868094 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.888220 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.901888 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.914868 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.928469 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.944476 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:50Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.944613 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.944740 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:50 crc kubenswrapper[4573]: E1203 08:38:50.944775 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:39:22.944752914 +0000 UTC m=+83.513132263 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:38:50 crc kubenswrapper[4573]: E1203 08:38:50.944851 4573 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:50 crc kubenswrapper[4573]: E1203 08:38:50.944919 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:39:22.944902498 +0000 UTC m=+83.513281767 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.956840 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.956874 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.956882 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.956894 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:50 crc kubenswrapper[4573]: I1203 08:38:50.956902 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:50Z","lastTransitionTime":"2025-12-03T08:38:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.029968 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.030013 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.030149 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.030264 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.045824 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.045871 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.045920 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046001 4573 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046036 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046082 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046091 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046106 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046118 4573 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046121 4573 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046102 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:39:23.046080217 +0000 UTC m=+83.614459556 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046223 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 08:39:23.04618463 +0000 UTC m=+83.614563929 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.046282 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 08:39:23.046262102 +0000 UTC m=+83.614641491 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.059518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.059556 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.059578 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.059593 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.059602 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.161371 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.161435 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.161453 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.161474 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.161490 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.263982 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.264022 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.264034 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.264078 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.264092 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.367108 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.367143 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.367153 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.367165 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.367174 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.469973 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.470013 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.470023 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.470040 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.470072 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.573116 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.573181 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.573200 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.573226 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.573244 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.676647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.676717 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.676738 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.676766 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.676789 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.716959 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/2.log" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.717792 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/1.log" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.721120 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3" exitCode=1 Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.721171 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.721209 4573 scope.go:117] "RemoveContainer" containerID="01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.722190 4573 scope.go:117] "RemoveContainer" containerID="bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3" Dec 03 08:38:51 crc kubenswrapper[4573]: E1203 08:38:51.722427 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.740300 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.758109 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01d8cc744a512a18e9b8c15e51a94b8614b7c55bcb5a857f54aad24e0109defc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.967375 5907 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 08:38:33.967566 5907 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:38:33.968341 5907 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 08:38:33.968367 5907 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 08:38:33.968394 5907 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 08:38:33.968399 5907 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 08:38:33.968403 5907 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 08:38:33.968412 5907 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 08:38:33.968417 5907 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 08:38:33.968425 5907 factory.go:656] Stopping watch factory\\\\nI1203 08:38:33.968426 5907 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 08:38:33.968436 5907 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 08:38:33.968436 5907 ovnkube.go:599] Stopped ovnkube\\\\nI1203 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.769287 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.779460 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.779499 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.779511 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.779531 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.779544 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.783213 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.795652 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.806835 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.821507 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.833845 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.846805 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.857880 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.868954 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.882679 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.882993 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.883137 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.883283 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.883395 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.883012 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.899067 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.912805 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.925312 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.934819 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.947567 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:51Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.985816 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.986165 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.986272 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.986364 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:51 crc kubenswrapper[4573]: I1203 08:38:51.986463 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:51Z","lastTransitionTime":"2025-12-03T08:38:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.029383 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.029688 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:52 crc kubenswrapper[4573]: E1203 08:38:52.029848 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:52 crc kubenswrapper[4573]: E1203 08:38:52.030211 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.088915 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.088968 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.088986 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.089008 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.089024 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.191141 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.191172 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.191181 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.191195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.191205 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.293590 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.293631 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.293644 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.293662 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.293674 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.396815 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.396857 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.396868 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.396886 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.396899 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.499919 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.499953 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.499963 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.499976 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.499987 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.602371 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.602420 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.602431 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.602444 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.602456 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.704735 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.705016 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.705025 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.705037 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.705066 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.725573 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/2.log" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.728770 4573 scope.go:117] "RemoveContainer" containerID="bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3" Dec 03 08:38:52 crc kubenswrapper[4573]: E1203 08:38:52.728966 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.749688 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.764864 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.777924 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.790568 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.803122 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.807796 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.807832 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.807861 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.807875 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.807884 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.815857 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.828661 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.840213 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.850526 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.861867 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.871280 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.879830 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.888652 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.898134 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.909525 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.909554 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.909562 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.909589 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.909599 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:52Z","lastTransitionTime":"2025-12-03T08:38:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.917811 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.928614 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:52 crc kubenswrapper[4573]: I1203 08:38:52.939907 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:52Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.011899 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.011949 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.011960 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.012004 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.012019 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.030326 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:53 crc kubenswrapper[4573]: E1203 08:38:53.030438 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.030339 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:53 crc kubenswrapper[4573]: E1203 08:38:53.030910 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.114681 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.114746 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.114763 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.114786 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.114803 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.218315 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.218374 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.218404 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.218428 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.218445 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.321269 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.321314 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.321327 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.321343 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.321356 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.424091 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.424124 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.424134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.424149 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.424160 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.526948 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.526988 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.526997 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.527012 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.527022 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.629708 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.629749 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.629761 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.629774 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.629785 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.731689 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.731748 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.731762 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.731780 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.731792 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.834606 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.834647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.834658 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.834674 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.834686 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.937553 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.937625 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.937635 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.937655 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:53 crc kubenswrapper[4573]: I1203 08:38:53.937668 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:53Z","lastTransitionTime":"2025-12-03T08:38:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.030241 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.030286 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:54 crc kubenswrapper[4573]: E1203 08:38:54.030390 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:54 crc kubenswrapper[4573]: E1203 08:38:54.030476 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.040387 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.040421 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.040433 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.040448 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.040461 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.143371 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.143430 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.143448 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.143483 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.143500 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.246000 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.246040 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.246071 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.246086 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.246098 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.349139 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.349182 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.349192 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.349206 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.349215 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.452019 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.452079 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.452092 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.452107 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.452120 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.554838 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.554879 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.554891 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.554909 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.554920 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.657676 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.657738 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.657748 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.657767 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.657780 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.761504 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.761558 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.761576 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.761596 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.761610 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.864395 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.864463 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.864474 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.864491 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.864501 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.967167 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.967228 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.967237 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.967269 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:54 crc kubenswrapper[4573]: I1203 08:38:54.967280 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:54Z","lastTransitionTime":"2025-12-03T08:38:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.030363 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.030411 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:55 crc kubenswrapper[4573]: E1203 08:38:55.030567 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:55 crc kubenswrapper[4573]: E1203 08:38:55.030741 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.070113 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.070164 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.070174 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.070196 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.070208 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.173965 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.174028 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.174078 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.174113 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.174136 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.276414 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.276468 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.276498 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.276518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.276533 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.378979 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.379025 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.379042 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.379101 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.379120 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.482278 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.482318 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.482332 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.482357 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.482372 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.585164 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.585207 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.585219 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.585235 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.585249 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.687483 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.687526 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.687535 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.687549 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.687559 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.790190 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.790240 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.790251 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.790267 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.790280 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.892601 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.892683 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.892693 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.892708 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.892718 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.995577 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.995618 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.995628 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.995641 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:55 crc kubenswrapper[4573]: I1203 08:38:55.995655 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:55Z","lastTransitionTime":"2025-12-03T08:38:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.030371 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.030370 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:56 crc kubenswrapper[4573]: E1203 08:38:56.030505 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:56 crc kubenswrapper[4573]: E1203 08:38:56.030660 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.097627 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.097661 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.097670 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.097684 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.097693 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.199821 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.199867 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.199879 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.199897 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.199908 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.302635 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.302687 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.302698 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.302715 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.302725 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.405777 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.405822 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.405835 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.405852 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.405863 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.508016 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.508066 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.508075 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.508092 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.508104 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.533634 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.533702 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.533734 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.533749 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.533760 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: E1203 08:38:56.547394 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:56Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.551038 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.551122 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.551140 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.551163 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.551181 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: E1203 08:38:56.562897 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:56Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.568727 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.568778 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.568794 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.568821 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.568835 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: E1203 08:38:56.586279 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:56Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.590157 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.590206 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.590218 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.590236 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.590248 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: E1203 08:38:56.602708 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:56Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.605907 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.605956 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.605968 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.605983 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.605994 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: E1203 08:38:56.619468 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:38:56Z is after 2025-08-24T17:21:41Z" Dec 03 08:38:56 crc kubenswrapper[4573]: E1203 08:38:56.619649 4573 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.621332 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.621418 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.621435 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.621454 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.621971 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.725589 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.725645 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.725654 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.725669 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.725678 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.827581 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.827610 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.827621 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.827634 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.827644 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.929752 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.929804 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.929818 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.929834 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:56 crc kubenswrapper[4573]: I1203 08:38:56.929845 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:56Z","lastTransitionTime":"2025-12-03T08:38:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.029320 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.029398 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:57 crc kubenswrapper[4573]: E1203 08:38:57.029527 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:57 crc kubenswrapper[4573]: E1203 08:38:57.029778 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.032848 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.032903 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.032917 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.032936 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.032949 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.135561 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.135608 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.135621 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.135637 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.135649 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.237604 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.237649 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.237663 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.237679 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.237692 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.340689 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.340767 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.340805 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.340839 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.340864 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.444405 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.444455 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.444464 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.444479 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.444488 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.547858 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.547915 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.547930 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.547952 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.547970 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.651195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.651259 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.651272 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.651293 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.651306 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.754570 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.754642 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.754660 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.754688 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.754706 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.858354 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.858406 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.858420 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.858437 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.858449 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.961308 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.961352 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.961362 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.961377 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:57 crc kubenswrapper[4573]: I1203 08:38:57.961389 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:57Z","lastTransitionTime":"2025-12-03T08:38:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.029596 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.029754 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:38:58 crc kubenswrapper[4573]: E1203 08:38:58.030101 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:38:58 crc kubenswrapper[4573]: E1203 08:38:58.030315 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.064571 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.064891 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.065136 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.065282 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.065448 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.168569 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.168856 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.168949 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.169107 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.169172 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.271491 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.271535 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.271546 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.271565 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.271576 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.375220 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.375286 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.375303 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.375326 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.375344 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.478484 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.478621 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.478635 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.478650 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.478663 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.580568 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.580610 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.580622 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.580639 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.580650 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.682912 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.682971 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.682989 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.683011 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.683028 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.784844 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.784901 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.784919 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.784941 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.784959 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.887939 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.887977 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.887986 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.888000 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.888010 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.991000 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.991324 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.991395 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.991463 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:58 crc kubenswrapper[4573]: I1203 08:38:58.991519 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:58Z","lastTransitionTime":"2025-12-03T08:38:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.030112 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.030168 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:38:59 crc kubenswrapper[4573]: E1203 08:38:59.030294 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:38:59 crc kubenswrapper[4573]: E1203 08:38:59.030346 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.094312 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.094348 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.094359 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.094375 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.094388 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.197118 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.197179 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.197191 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.197216 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.197232 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.299912 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.299983 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.300000 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.300024 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.300041 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.402897 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.402963 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.402980 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.403010 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.403028 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.505799 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.505860 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.505877 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.505901 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.505919 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.608518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.608851 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.608875 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.608904 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.608926 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.712810 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.712888 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.712899 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.712921 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.712937 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.816544 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.816600 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.816616 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.816638 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.816655 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.920434 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.920499 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.920521 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.920555 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:38:59 crc kubenswrapper[4573]: I1203 08:38:59.920577 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:38:59Z","lastTransitionTime":"2025-12-03T08:38:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.023559 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.023604 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.023623 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.023647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.023660 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.029732 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.029797 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:00 crc kubenswrapper[4573]: E1203 08:39:00.029931 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:00 crc kubenswrapper[4573]: E1203 08:39:00.029988 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.048425 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.065796 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.079434 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.095981 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.111868 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.130760 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.130816 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.130854 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.130868 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.130887 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.130903 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.142957 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.155816 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.177236 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.193780 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.213460 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.227842 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.233983 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.234029 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.234041 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.234080 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.234094 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.241939 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.258306 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.274593 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.289633 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.301791 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:00Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.337021 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.337125 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.337142 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.337159 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.337170 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.439863 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.439913 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.439923 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.439943 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.439955 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.543401 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.543457 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.543473 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.543511 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.543560 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.646694 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.646750 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.646762 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.646784 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.646808 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.748847 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.748885 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.748895 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.748913 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.748925 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.851802 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.851841 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.851851 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.851869 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.851883 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.954440 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.954498 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.954511 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.954533 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:00 crc kubenswrapper[4573]: I1203 08:39:00.954549 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:00Z","lastTransitionTime":"2025-12-03T08:39:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.029743 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.029832 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:01 crc kubenswrapper[4573]: E1203 08:39:01.029968 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:01 crc kubenswrapper[4573]: E1203 08:39:01.030309 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.057960 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.058019 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.058032 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.058064 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.058078 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.160848 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.160889 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.160905 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.160921 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.160932 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.263218 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.263266 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.263277 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.263297 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.263314 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.366121 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.366160 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.366170 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.366187 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.366199 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.469317 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.469374 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.469385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.469406 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.469420 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.572606 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.572665 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.572676 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.572695 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.572707 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.675448 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.675540 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.675557 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.675579 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.675593 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.777724 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.777774 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.777782 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.777799 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.777810 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.880581 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.880624 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.880633 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.880650 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.880666 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.983388 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.983454 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.983466 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.983485 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:01 crc kubenswrapper[4573]: I1203 08:39:01.983526 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:01Z","lastTransitionTime":"2025-12-03T08:39:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.030095 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.030175 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:02 crc kubenswrapper[4573]: E1203 08:39:02.030274 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:02 crc kubenswrapper[4573]: E1203 08:39:02.030378 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.086276 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.086312 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.086322 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.086342 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.086353 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.189550 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.189588 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.189601 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.189617 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.189626 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.292541 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.292622 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.292685 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.292705 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.292722 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.396079 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.396125 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.396140 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.396161 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.396178 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.499639 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.499693 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.499707 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.499730 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.499746 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.602507 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.602549 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.602562 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.602585 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.602600 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.705419 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.705453 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.705461 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.705474 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.705501 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.808230 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.808299 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.808314 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.808332 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.808347 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.911762 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.911814 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.911825 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.911842 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:02 crc kubenswrapper[4573]: I1203 08:39:02.911853 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:02Z","lastTransitionTime":"2025-12-03T08:39:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.015188 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.015235 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.015244 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.015261 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.015274 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.030275 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.030317 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:03 crc kubenswrapper[4573]: E1203 08:39:03.030528 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:03 crc kubenswrapper[4573]: E1203 08:39:03.030629 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.118686 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.118788 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.118807 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.118835 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.118852 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.222609 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.222654 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.222666 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.222683 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.222697 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.325663 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.325730 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.325744 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.325767 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.325782 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.428613 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.428656 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.428665 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.428681 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.428691 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.532248 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.532300 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.532311 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.532331 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.532345 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.634814 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.634852 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.634864 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.634885 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.634899 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.737827 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.737879 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.737898 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.737923 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.737940 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.841825 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.841877 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.841894 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.841921 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.841938 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.944572 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.944616 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.944628 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.944642 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:03 crc kubenswrapper[4573]: I1203 08:39:03.944652 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:03Z","lastTransitionTime":"2025-12-03T08:39:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.029578 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.029707 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:04 crc kubenswrapper[4573]: E1203 08:39:04.029799 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:04 crc kubenswrapper[4573]: E1203 08:39:04.029972 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.030690 4573 scope.go:117] "RemoveContainer" containerID="bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3" Dec 03 08:39:04 crc kubenswrapper[4573]: E1203 08:39:04.030964 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.046674 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.046720 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.046731 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.046754 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.046765 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.149766 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.150104 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.150121 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.150140 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.150151 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.252981 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.253067 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.253080 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.253104 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.253121 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.356289 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.356356 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.356376 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.356406 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.356424 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.459772 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.459809 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.459819 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.459837 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.459848 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.562736 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.562775 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.562787 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.562804 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.562813 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.665219 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.665256 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.665265 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.665281 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.665292 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.767733 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.768134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.768217 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.768326 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.768408 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.871446 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.871484 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.871497 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.871521 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.871534 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.974565 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.974632 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.974645 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.974659 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:04 crc kubenswrapper[4573]: I1203 08:39:04.974669 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:04Z","lastTransitionTime":"2025-12-03T08:39:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.030403 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:05 crc kubenswrapper[4573]: E1203 08:39:05.030550 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.030769 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:05 crc kubenswrapper[4573]: E1203 08:39:05.030850 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.077698 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.077754 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.077763 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.077780 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.077791 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.180800 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.180863 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.180875 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.180896 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.180914 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.283092 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.283145 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.283161 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.283183 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.283196 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.386259 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.386576 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.386646 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.386725 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.386787 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.488952 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.489289 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.489357 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.489437 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.489507 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.591904 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.591958 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.591975 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.591998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.592015 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.694465 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.694512 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.694526 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.694542 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.694554 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.796831 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.796869 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.796882 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.796899 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.796912 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.899137 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.899486 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.899577 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.899685 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:05 crc kubenswrapper[4573]: I1203 08:39:05.899780 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:05Z","lastTransitionTime":"2025-12-03T08:39:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.002850 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.002897 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.002907 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.002921 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.002932 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.029562 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.029604 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.029726 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.029810 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.106233 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.106273 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.106289 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.106312 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.106328 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.208797 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.208842 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.208851 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.208864 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.208874 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.310894 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.311199 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.311288 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.311370 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.311452 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.415299 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.415366 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.415375 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.415409 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.415422 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.511220 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.511440 4573 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.511568 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs podName:bb7f6603-0f41-4e0f-b77e-47dbcc676aaf nodeName:}" failed. No retries permitted until 2025-12-03 08:39:38.511540554 +0000 UTC m=+99.079919873 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs") pod "network-metrics-daemon-kcphv" (UID: "bb7f6603-0f41-4e0f-b77e-47dbcc676aaf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.517322 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.517375 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.517385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.517399 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.517408 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.619527 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.619590 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.619606 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.619629 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.619648 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.721876 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.721924 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.721937 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.721958 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.721970 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.735003 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.735040 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.735063 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.735078 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.735089 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.747426 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:06Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.751984 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.752180 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.752285 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.752400 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.752499 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.765767 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:06Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.770684 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.770735 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.770747 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.770766 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.770780 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.785239 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:06Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.789545 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.789591 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.789603 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.789618 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.789628 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.805908 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:06Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.809621 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.809670 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.809680 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.809693 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.809703 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.822310 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:06Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:06 crc kubenswrapper[4573]: E1203 08:39:06.822431 4573 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.823801 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.823827 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.823838 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.823858 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.823870 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.927376 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.927440 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.927459 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.927482 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:06 crc kubenswrapper[4573]: I1203 08:39:06.927497 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:06Z","lastTransitionTime":"2025-12-03T08:39:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.029703 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:07 crc kubenswrapper[4573]: E1203 08:39:07.029821 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.029718 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:07 crc kubenswrapper[4573]: E1203 08:39:07.030019 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.030142 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.030175 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.030187 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.030203 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.030216 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.133266 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.133322 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.133334 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.133358 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.133372 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.236099 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.236157 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.236172 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.236190 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.236204 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.339568 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.339625 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.339637 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.339658 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.339670 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.442702 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.442761 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.442775 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.442797 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.442812 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.545187 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.545236 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.545250 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.545268 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.545283 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.648299 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.648345 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.648359 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.648380 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.648393 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.751424 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.751815 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.751923 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.752043 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.752202 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.855297 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.855383 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.855405 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.855439 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.855459 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.958182 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.958235 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.958249 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.958271 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:07 crc kubenswrapper[4573]: I1203 08:39:07.958284 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:07Z","lastTransitionTime":"2025-12-03T08:39:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.030340 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.030352 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:08 crc kubenswrapper[4573]: E1203 08:39:08.030570 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:08 crc kubenswrapper[4573]: E1203 08:39:08.030659 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.061009 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.061060 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.061070 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.061085 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.061097 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.163824 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.163874 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.163887 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.163905 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.163918 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.266958 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.267076 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.267100 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.267128 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.267180 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.369873 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.369928 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.369940 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.369964 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.369978 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.473102 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.473157 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.473168 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.473190 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.473202 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.576300 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.576353 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.576366 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.576385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.576397 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.679475 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.679541 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.679555 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.679578 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.679594 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.782301 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.782365 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.782384 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.782400 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.782412 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.886067 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.886143 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.886156 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.886177 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.886191 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.989333 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.989403 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.989414 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.989437 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:08 crc kubenswrapper[4573]: I1203 08:39:08.989447 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:08Z","lastTransitionTime":"2025-12-03T08:39:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.030336 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.030398 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:09 crc kubenswrapper[4573]: E1203 08:39:09.030878 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:09 crc kubenswrapper[4573]: E1203 08:39:09.030941 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.091560 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.091652 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.091671 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.091699 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.091718 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.194168 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.194508 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.194629 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.194752 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.194841 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.298290 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.298625 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.298730 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.298902 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.299029 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.402855 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.403280 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.404039 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.404370 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.404477 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.507998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.508077 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.508088 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.508188 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.508275 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.610998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.611094 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.611106 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.611126 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.611138 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.714541 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.714631 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.714650 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.714683 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.714704 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.783759 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/0.log" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.783949 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c98ef96-0e91-49e3-88d5-c95dd6f52c5d" containerID="296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0" exitCode=1 Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.784017 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ch24s" event={"ID":"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d","Type":"ContainerDied","Data":"296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.784821 4573 scope.go:117] "RemoveContainer" containerID="296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.808352 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.818128 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.818173 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.818182 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.818201 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.818212 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.823213 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.838269 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.852550 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.865664 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.876929 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.890093 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.913417 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.920417 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.920696 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.920823 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.920963 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.921092 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:09Z","lastTransitionTime":"2025-12-03T08:39:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.928711 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.946183 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"2025-12-03T08:38:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f\\\\n2025-12-03T08:38:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f to /host/opt/cni/bin/\\\\n2025-12-03T08:38:24Z [verbose] multus-daemon started\\\\n2025-12-03T08:38:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T08:39:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.961072 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.976738 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:09 crc kubenswrapper[4573]: I1203 08:39:09.990427 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:09Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.005926 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.023683 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.023739 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.023751 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.023825 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.023839 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.029375 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:10 crc kubenswrapper[4573]: E1203 08:39:10.029495 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.029733 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:10 crc kubenswrapper[4573]: E1203 08:39:10.029925 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.033774 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.052334 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.069246 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.082957 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.103814 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.129286 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.129323 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.129334 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.129349 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.129360 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.143182 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.172282 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.192554 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.225410 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.231724 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.231788 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.231800 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.231841 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.231854 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.250065 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.267731 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.286294 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.305154 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.317335 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.331010 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.335444 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.335486 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.335498 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.335516 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.335527 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.344324 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.354666 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.368892 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"2025-12-03T08:38:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f\\\\n2025-12-03T08:38:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f to /host/opt/cni/bin/\\\\n2025-12-03T08:38:24Z [verbose] multus-daemon started\\\\n2025-12-03T08:38:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T08:39:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.379450 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.392308 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.568610 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.568651 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.568665 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.568687 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.568699 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.671119 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.671181 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.671192 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.671209 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.671220 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.775371 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.775411 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.775421 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.775440 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.775453 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.788846 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/0.log" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.788906 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ch24s" event={"ID":"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d","Type":"ContainerStarted","Data":"35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.805730 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.820104 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.835268 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.849707 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.878309 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.878358 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.878370 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.878389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.878401 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.901681 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.916033 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.929064 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.940893 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.955850 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.973393 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.981231 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.981273 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.981289 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.981311 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.981328 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:10Z","lastTransitionTime":"2025-12-03T08:39:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:10 crc kubenswrapper[4573]: I1203 08:39:10.989570 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:10Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.001908 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"2025-12-03T08:38:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f\\\\n2025-12-03T08:38:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f to /host/opt/cni/bin/\\\\n2025-12-03T08:38:24Z [verbose] multus-daemon started\\\\n2025-12-03T08:38:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T08:39:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:39:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:11Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.016762 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:11Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.029073 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:11Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.029571 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.029592 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:11 crc kubenswrapper[4573]: E1203 08:39:11.029788 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:11 crc kubenswrapper[4573]: E1203 08:39:11.029997 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.050719 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:11Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.068807 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:11Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.084025 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.084082 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.084094 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.084112 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.084125 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.085754 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:11Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.187191 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.187228 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.187240 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.187258 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.187286 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.290394 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.290444 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.290456 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.290477 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.290491 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.394101 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.394164 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.394177 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.394198 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.394211 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.498383 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.498421 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.498434 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.498453 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.498464 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.602655 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.602715 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.602728 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.602750 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.602793 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.707072 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.707127 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.707141 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.707165 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.707182 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.809842 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.809893 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.809909 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.809929 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.809942 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.913271 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.913345 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.913357 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.913378 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:11 crc kubenswrapper[4573]: I1203 08:39:11.913391 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:11Z","lastTransitionTime":"2025-12-03T08:39:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.016700 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.016758 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.016769 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.016788 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.016801 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.030252 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.030403 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:12 crc kubenswrapper[4573]: E1203 08:39:12.030529 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:12 crc kubenswrapper[4573]: E1203 08:39:12.030795 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.046286 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.119833 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.119879 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.119908 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.119931 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.119951 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.222823 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.222876 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.222903 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.222928 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.222939 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.325810 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.325885 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.325909 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.325944 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.325967 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.428555 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.428607 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.428619 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.428640 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.428654 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.531505 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.531562 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.531577 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.531597 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.531613 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.634320 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.634406 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.634459 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.634478 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.634493 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.737493 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.737543 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.737553 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.737571 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.737586 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.840426 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.840476 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.840490 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.840513 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.840525 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.944171 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.944227 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.944242 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.944263 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:12 crc kubenswrapper[4573]: I1203 08:39:12.944279 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:12Z","lastTransitionTime":"2025-12-03T08:39:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.029337 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:13 crc kubenswrapper[4573]: E1203 08:39:13.029484 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.029547 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:13 crc kubenswrapper[4573]: E1203 08:39:13.029809 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.047995 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.048084 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.048098 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.048118 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.048130 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.151013 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.151104 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.151124 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.151147 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.151164 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.254246 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.254304 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.254315 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.254336 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.254355 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.356919 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.356955 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.356964 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.356978 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.356988 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.460337 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.460540 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.460722 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.460857 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.460900 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.564552 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.564606 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.564620 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.564638 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.564651 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.667876 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.667925 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.667936 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.667953 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.667964 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.771995 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.772078 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.772093 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.772111 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.772123 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.874459 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.874548 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.874561 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.874577 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.874589 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.978501 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.978555 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.978571 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.978595 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:13 crc kubenswrapper[4573]: I1203 08:39:13.978616 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:13Z","lastTransitionTime":"2025-12-03T08:39:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.030394 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:14 crc kubenswrapper[4573]: E1203 08:39:14.030536 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.030394 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:14 crc kubenswrapper[4573]: E1203 08:39:14.030752 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.082263 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.082351 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.082371 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.082407 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.082431 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.185250 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.185326 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.185347 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.185373 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.185390 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.289757 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.289848 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.289867 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.289892 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.289910 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.393168 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.393243 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.393257 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.393277 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.393290 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.496273 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.496338 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.496355 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.496385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.496404 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.599585 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.599635 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.599653 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.599675 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.599692 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.702659 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.702692 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.702700 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.702713 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.702722 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.804546 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.804588 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.804599 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.804616 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.804628 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.907460 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.907514 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.907524 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.907542 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:14 crc kubenswrapper[4573]: I1203 08:39:14.907881 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:14Z","lastTransitionTime":"2025-12-03T08:39:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.009988 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.010019 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.010027 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.010039 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.010069 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.029673 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:15 crc kubenswrapper[4573]: E1203 08:39:15.029768 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.029905 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:15 crc kubenswrapper[4573]: E1203 08:39:15.029965 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.112544 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.112582 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.112596 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.112612 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.112624 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.217692 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.217771 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.217814 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.217843 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.217857 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.321146 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.321217 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.321244 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.321269 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.321284 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.424949 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.425012 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.425027 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.425081 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.425106 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.528034 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.528115 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.528124 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.528137 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.528150 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.630166 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.630202 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.630214 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.630227 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.630242 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.733595 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.733657 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.733672 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.733689 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.733702 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.836191 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.836247 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.836259 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.836281 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.836293 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.939166 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.939245 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.939263 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.939293 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:15 crc kubenswrapper[4573]: I1203 08:39:15.939313 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:15Z","lastTransitionTime":"2025-12-03T08:39:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.030103 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.030174 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:16 crc kubenswrapper[4573]: E1203 08:39:16.030307 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:16 crc kubenswrapper[4573]: E1203 08:39:16.030404 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.042194 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.042255 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.042269 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.042291 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.042304 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.145336 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.145376 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.145386 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.145407 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.145421 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.249373 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.249440 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.249452 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.249476 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.249489 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.354450 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.354507 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.354552 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.354575 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.354593 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.458162 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.458238 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.458258 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.458289 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.458309 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.561079 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.561161 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.561186 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.561217 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.561240 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.663716 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.663768 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.663786 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.663811 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.663828 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.767400 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.767482 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.767507 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.767538 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.767579 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.870514 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.870585 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.870607 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.870638 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.870660 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.974604 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.974657 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.974733 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.974762 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:16 crc kubenswrapper[4573]: I1203 08:39:16.974783 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:16Z","lastTransitionTime":"2025-12-03T08:39:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.029951 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.030013 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:17 crc kubenswrapper[4573]: E1203 08:39:17.030261 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:17 crc kubenswrapper[4573]: E1203 08:39:17.030311 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.061497 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.061566 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.061578 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.061607 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.061623 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: E1203 08:39:17.085643 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:17Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.090318 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.090368 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.090381 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.090401 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.090442 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: E1203 08:39:17.104239 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:17Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.108425 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.108502 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.108517 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.108564 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.108580 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: E1203 08:39:17.123208 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:17Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.128142 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.128182 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.128195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.128216 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.128230 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: E1203 08:39:17.143392 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:17Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.148797 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.148871 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.148916 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.148937 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.148951 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: E1203 08:39:17.166187 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:17Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:17 crc kubenswrapper[4573]: E1203 08:39:17.166350 4573 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.169088 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.169123 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.169134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.169151 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.169165 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.272584 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.272625 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.272638 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.272658 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.272674 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.375837 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.375888 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.375900 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.375920 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.375934 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.479252 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.479324 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.479339 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.479367 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.479385 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.582590 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.582653 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.582669 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.582689 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.582703 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.701650 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.701701 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.701714 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.701734 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.701749 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.829891 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.829947 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.829958 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.829976 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.829989 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.933593 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.933639 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.933649 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.933671 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:17 crc kubenswrapper[4573]: I1203 08:39:17.933682 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:17Z","lastTransitionTime":"2025-12-03T08:39:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.031384 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.031459 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:18 crc kubenswrapper[4573]: E1203 08:39:18.031908 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:18 crc kubenswrapper[4573]: E1203 08:39:18.031990 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.032174 4573 scope.go:117] "RemoveContainer" containerID="bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.036275 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.036342 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.036360 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.036386 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.036400 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.140319 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.140976 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.141010 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.141041 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.141110 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.244959 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.245015 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.245029 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.245076 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.245099 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.348120 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.348193 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.348208 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.348254 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.348269 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.452038 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.452119 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.452168 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.452197 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.452215 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.555932 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.555998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.556012 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.556064 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.556078 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.658586 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.658649 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.658675 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.658715 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.658739 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.761779 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.761835 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.761846 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.761864 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.761876 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.833802 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/2.log" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.836759 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.837425 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.852454 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.873545 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"2025-12-03T08:38:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f\\\\n2025-12-03T08:38:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f to /host/opt/cni/bin/\\\\n2025-12-03T08:38:24Z [verbose] multus-daemon started\\\\n2025-12-03T08:38:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T08:39:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:39:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.886925 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.901546 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.917821 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.935468 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.951838 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.955153 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.955205 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.955242 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.955266 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.955280 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:18Z","lastTransitionTime":"2025-12-03T08:39:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.977105 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:39:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:18 crc kubenswrapper[4573]: I1203 08:39:18.989415 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:18Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.007139 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.020977 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.034019 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.047736 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.056622 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.056666 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:19 crc kubenswrapper[4573]: E1203 08:39:19.056744 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:19 crc kubenswrapper[4573]: E1203 08:39:19.056864 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.058704 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.058735 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.058745 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.058761 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.058773 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.061197 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.087732 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.111539 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.126115 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3790c07f-51b1-4fbb-962b-33b614da7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2773e32404b31160beec1d08cef6235ae7c2909d701b7fbcfce92152b2d4fafe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ee9b66a1bc72eda0261abd5ee7d063582d9ac4127c87dd20f5c3a5684ca8cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ee9b66a1bc72eda0261abd5ee7d063582d9ac4127c87dd20f5c3a5684ca8cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.153520 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:19Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.162506 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.162564 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.162574 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.162593 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.162607 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.267358 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.267396 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.267406 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.267424 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.267435 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.374553 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.374617 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.374630 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.374673 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.374690 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.478122 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.478180 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.478193 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.478215 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.478232 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.584307 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.584377 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.584393 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.584415 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.584427 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.687488 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.687549 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.687563 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.687586 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.687600 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.790884 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.790938 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.790948 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.790972 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.790984 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.893546 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.893608 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.893625 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.893647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.893661 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.996379 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.996435 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.996447 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.996468 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:19 crc kubenswrapper[4573]: I1203 08:39:19.996487 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:19Z","lastTransitionTime":"2025-12-03T08:39:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.030301 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.030301 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:20 crc kubenswrapper[4573]: E1203 08:39:20.030645 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:20 crc kubenswrapper[4573]: E1203 08:39:20.030789 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.046778 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.048378 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.073549 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:39:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.099170 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.099742 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.099790 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.099811 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.099835 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.099852 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.133322 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.149296 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.163922 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.175022 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3790c07f-51b1-4fbb-962b-33b614da7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2773e32404b31160beec1d08cef6235ae7c2909d701b7fbcfce92152b2d4fafe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ee9b66a1bc72eda0261abd5ee7d063582d9ac4127c87dd20f5c3a5684ca8cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ee9b66a1bc72eda0261abd5ee7d063582d9ac4127c87dd20f5c3a5684ca8cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.192083 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.202715 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.202776 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.202792 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.202816 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.202845 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.210616 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.225636 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.241222 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.255570 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.267409 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.285156 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.305769 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.305824 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.305835 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.305853 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.305865 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.305988 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.321993 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.338543 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"2025-12-03T08:38:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f\\\\n2025-12-03T08:38:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f to /host/opt/cni/bin/\\\\n2025-12-03T08:38:24Z [verbose] multus-daemon started\\\\n2025-12-03T08:38:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T08:39:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:39:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.351812 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.409336 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.409412 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.409427 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.409445 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.409460 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.512580 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.512655 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.512674 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.512707 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.512729 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.615440 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.615487 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.615501 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.615518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.615531 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.717982 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.718068 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.718083 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.718103 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.718116 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.820589 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.820653 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.820666 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.820689 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.820703 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.846565 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/3.log" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.847255 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/2.log" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.850554 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" exitCode=1 Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.850656 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.850751 4573 scope.go:117] "RemoveContainer" containerID="bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.852154 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:39:20 crc kubenswrapper[4573]: E1203 08:39:20.852366 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.872500 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ea145c11e10f75d928bfb6b0fa7fd85099b4a38704e106491c005a7260365619\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733401790dc9b2a0e0725870c49e13f1bdd29d6473f251122304efb496276294\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.886745 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kcphv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vchs5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:34Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kcphv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.901660 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3790c07f-51b1-4fbb-962b-33b614da7e94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2773e32404b31160beec1d08cef6235ae7c2909d701b7fbcfce92152b2d4fafe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ee9b66a1bc72eda0261abd5ee7d063582d9ac4127c87dd20f5c3a5684ca8cc9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ee9b66a1bc72eda0261abd5ee7d063582d9ac4127c87dd20f5c3a5684ca8cc9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.918252 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6679bb39-7c2c-4518-bf40-1a1abf9ebc66\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 08:38:13.695090 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 08:38:13.696105 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2370552256/tls.crt::/tmp/serving-cert-2370552256/tls.key\\\\\\\"\\\\nI1203 08:38:19.046118 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 08:38:19.067494 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 08:38:19.067677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 08:38:19.067717 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 08:38:19.067726 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 08:38:19.146267 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 08:38:19.146302 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 08:38:19.146316 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 08:38:19.146319 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 08:38:19.146323 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 08:38:19.146327 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 08:38:19.146529 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 08:38:19.167533 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.923248 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.923277 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.923289 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.923308 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.923323 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:20Z","lastTransitionTime":"2025-12-03T08:39:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.930549 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ed097053-b08a-4e04-a1d9-91947bdf7dbe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3651610e436a9e8042dae9ee3e01b58b0e836ef2444865a08e6e89af54b730f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a285015603d50d9f7d7c29befbd9986dfcfbfb9abbc2d933e3c1e2823b8af82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://567f6c87336254eabd32ee4d335dc700499bffa7b37d691f9db21a42baa77bea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://989b5a83891df12e27ff5a8b60d7a9e907fe88a282e70cd2d70d136087e971c3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.945166 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.959486 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.978315 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:20 crc kubenswrapper[4573]: I1203 08:39:20.994714 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:20Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.015668 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.026499 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.026556 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.026571 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.026590 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.026823 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.029612 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:21 crc kubenswrapper[4573]: E1203 08:39:21.029734 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.029612 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:21 crc kubenswrapper[4573]: E1203 08:39:21.029851 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.032084 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:18Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.050251 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://863c8c10513182e42fc01dccd1e23296d8f5023a391560d98580ffc1f1988908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.065604 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"2025-12-03T08:38:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f\\\\n2025-12-03T08:38:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f to /host/opt/cni/bin/\\\\n2025-12-03T08:38:24Z [verbose] multus-daemon started\\\\n2025-12-03T08:38:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T08:39:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:39:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.077996 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.097086 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8faa3b55-12bf-445e-be3c-878f1b3db392\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6abdb217c16638eac968f2ca5423deacf945c63f23308388f50b6a7842dcbf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4ea1ae440959c88c05edb78b0ef1031186e899821b71e589a2e91776f355f2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://58c67c54dbacf5ef99060c0aa0fc9afb05148989645cb74acb236a3107b3d9d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a686c18023ff96fec4885ee571151253ff025601b7afb973b13c54885dff1cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7365e4e47cd6dc02ccee8b96fadedc0a9779682191b812efe58606f4d371a82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61424a4f23000a56999960cbdb89e67e37b0bd43a74219bcf88fb6a50dff1da4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61424a4f23000a56999960cbdb89e67e37b0bd43a74219bcf88fb6a50dff1da4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d21665c9762f4c5794e20734889f39231981a2c0c677e8090ad2980e85e1755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d21665c9762f4c5794e20734889f39231981a2c0c677e8090ad2980e85e1755\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://87742c98718b8ff664bb42a7abb281b1465df3e245c35b941ec825fb396547f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87742c98718b8ff664bb42a7abb281b1465df3e245c35b941ec825fb396547f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.109846 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-blvnm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"56192d0e-1b05-4a3d-b176-b3d27de36527\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3379f2568f320e9b3c54d0c2b9a3d3bf6dd8d3fc4f653f1f5439dc44bc210846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vsm9q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-blvnm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.128853 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.128746 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48186949-d112-4fb7-aa60-026d0176c73e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfdb8b5897f7b40f381bdd200fe807dee1d14780d43d3f2d8bc87962df1319e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:38:51Z\\\",\\\"message\\\":\\\"(nil)}}\\\\nI1203 08:38:51.212646 6099 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1203 08:38:51.213242 6099 services_controller.go:453] Built service default/kubernetes template LB for network=default: []services.LB{}\\\\nI1203 08:38:51.213244 6099 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1203 08:38:51.213251 6099 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1203 08:38:51.213252 6099 services_controller.go:454] Service default/kubernetes for network=default has 0 cluster-wide, 1 per-node configs, 0 template configs, making 0 (cluster) 2 (per node) and 0 (template) load balancers\\\\nF1203 08:38:51.212759 6099 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:39:20Z\\\",\\\"message\\\":\\\"work-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 08:39:20.393819 6429 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 08:39:20.395084 6429 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:39:20.395262 6429 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:39:20.395493 6429 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:39:20.395883 6429 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:39:20.396134 6429 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 08:39:20.418905 6429 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1203 08:39:20.418948 6429 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1203 08:39:20.419006 6429 ovnkube.go:599] Stopped ovnkube\\\\nI1203 08:39:20.419162 6429 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 08:39:20.419405 6429 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:39:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fssxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x466z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.128942 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.128956 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.128979 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.128995 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.143229 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://753f66ef4ea59d15cc1d01e6f6a63279b4247bf12d8a4c4ed66e1053420ce192\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.161097 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60e19539-4df1-4909-b8ee-c6ac0fcfc52d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c5ffb6579a8784c3349122bb36beab883c410dd2ae2fe22dbdd7c11d007d5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://821cdaba54759646402570da87a3ce9741eb7cf885bfce87c3444bb58e3d2e1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7622a3ac7e3fb0f033fb13f82c8aa355057f4f590481c172bf46af139438c7d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6252147ae2a87843e5f15e450c980a65a0f501c6cf3c1d4b3f53e0fa72f4733\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://914bf46b7a51b5dc33c6780e1b9ff94298e636b535965ca6992d273e54bf2e43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4940292b8c33a24cea34e75f8da821261ab6b997762ea2f6f53e14e4eda7fa93\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d230fb96972be9e021e927d66cc9aa021cc337a66d103018eb85907fd4df1f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T08:38:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pj9xr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfcvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:21Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.231569 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.231616 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.231628 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.231646 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.231657 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.334652 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.334724 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.334737 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.334758 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.334771 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.438380 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.438445 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.438465 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.438492 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.438514 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.541309 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.541365 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.541376 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.541396 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.541410 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.644584 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.644647 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.644658 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.644680 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.644693 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.747763 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.747822 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.747836 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.747859 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.747876 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.850932 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.850999 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.851030 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.851089 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.851108 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.855681 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/3.log" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.955118 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.955169 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.955185 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.955211 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:21 crc kubenswrapper[4573]: I1203 08:39:21.955231 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:21Z","lastTransitionTime":"2025-12-03T08:39:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.029507 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.029543 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:22 crc kubenswrapper[4573]: E1203 08:39:22.029685 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:22 crc kubenswrapper[4573]: E1203 08:39:22.029919 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.058034 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.058100 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.058115 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.058134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.058148 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.160955 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.160998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.161011 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.161033 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.161060 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.263985 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.264073 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.264098 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.264131 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.264153 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.367125 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.367179 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.367193 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.367215 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.367228 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.470555 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.470608 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.470618 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.470640 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.470653 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.573973 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.574019 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.574031 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.574071 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.574088 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.677258 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.677307 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.677318 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.677337 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.677349 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.780754 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.780804 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.780818 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.780838 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.780852 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.882882 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.882930 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.882945 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.882968 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.882980 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.985968 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.986037 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.986075 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.986102 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:22 crc kubenswrapper[4573]: I1203 08:39:22.986121 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:22Z","lastTransitionTime":"2025-12-03T08:39:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.029384 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.029384 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.029581 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.029657 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.039693 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.039830 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.040070 4573 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.040094 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.040000198 +0000 UTC m=+147.608379497 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.040262 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.040236195 +0000 UTC m=+147.608615644 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.089675 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.089718 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.089728 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.089744 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.089755 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.140777 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.140850 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.140894 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141109 4573 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141190 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.14116659 +0000 UTC m=+147.709545869 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141207 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141245 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141302 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141311 4573 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141355 4573 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141357 4573 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141454 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.141427317 +0000 UTC m=+147.709806586 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:39:23 crc kubenswrapper[4573]: E1203 08:39:23.141512 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.141479898 +0000 UTC m=+147.709859167 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.193134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.193275 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.193307 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.193348 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.193381 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.297090 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.297152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.297164 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.297185 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.297198 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.401291 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.401343 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.401355 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.401377 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.401392 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.505035 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.505175 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.505191 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.505219 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.505236 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.608830 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.608878 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.608889 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.608912 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.608925 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.712819 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.712897 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.712913 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.712939 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.712954 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.815936 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.816011 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.816023 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.816058 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.816072 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.918584 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.918657 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.918670 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.918694 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:23 crc kubenswrapper[4573]: I1203 08:39:23.918708 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:23Z","lastTransitionTime":"2025-12-03T08:39:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.021199 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.021261 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.021272 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.021294 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.021308 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.029315 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.029339 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:24 crc kubenswrapper[4573]: E1203 08:39:24.029451 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:24 crc kubenswrapper[4573]: E1203 08:39:24.029545 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.124543 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.124589 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.124608 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.124627 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.124646 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.228670 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.228732 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.228743 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.228761 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.228773 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.331671 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.331718 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.331730 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.331752 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.331768 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.435021 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.435095 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.435106 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.435126 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.435137 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.538121 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.538183 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.538197 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.538220 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.538235 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.641473 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.641580 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.641599 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.641624 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.641640 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.746152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.746504 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.746518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.746541 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.746553 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.850106 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.850151 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.850162 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.850178 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.850190 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.953160 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.953212 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.953225 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.953247 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:24 crc kubenswrapper[4573]: I1203 08:39:24.953263 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:24Z","lastTransitionTime":"2025-12-03T08:39:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.029929 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.030076 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:25 crc kubenswrapper[4573]: E1203 08:39:25.030118 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:25 crc kubenswrapper[4573]: E1203 08:39:25.030268 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.055839 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.055892 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.055903 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.055922 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.055935 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.158400 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.158465 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.158480 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.158503 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.158516 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.261422 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.261467 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.261476 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.261496 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.261507 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.364260 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.364317 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.364332 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.364354 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.364367 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.467744 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.467799 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.467813 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.467838 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.467849 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.570676 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.570730 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.570740 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.570760 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.570775 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.673137 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.673187 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.673196 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.673218 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.673228 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.775947 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.775995 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.776006 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.776025 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.776038 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.884409 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.884450 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.884466 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.884485 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.884498 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.986737 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.986775 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.986812 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.986836 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:25 crc kubenswrapper[4573]: I1203 08:39:25.986849 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:25Z","lastTransitionTime":"2025-12-03T08:39:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.030400 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.030524 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:26 crc kubenswrapper[4573]: E1203 08:39:26.030654 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:26 crc kubenswrapper[4573]: E1203 08:39:26.030816 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.089146 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.089195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.089207 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.089227 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.089239 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.192319 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.192373 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.192387 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.192407 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.192419 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.295445 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.295503 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.295522 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.295544 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.295562 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.398670 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.398719 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.398749 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.398777 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.398795 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.501729 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.501790 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.501800 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.501817 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.501829 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.604481 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.604653 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.604675 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.604695 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.604711 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.707399 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.707445 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.707455 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.707472 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.707482 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.809435 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.809490 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.809502 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.809518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.809527 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.913388 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.913479 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.913502 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.913534 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:26 crc kubenswrapper[4573]: I1203 08:39:26.913558 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:26Z","lastTransitionTime":"2025-12-03T08:39:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.017521 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.017611 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.017695 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.017727 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.017752 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.029917 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.029938 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:27 crc kubenswrapper[4573]: E1203 08:39:27.030147 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:27 crc kubenswrapper[4573]: E1203 08:39:27.030440 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.121136 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.121202 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.121221 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.121249 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.121269 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.223812 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.223894 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.223915 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.223948 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.223969 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.327711 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.327761 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.327773 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.327793 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.327807 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.345385 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.345438 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.345457 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.345478 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.345490 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: E1203 08:39:27.363927 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.369499 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.369552 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.369566 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.369586 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.369599 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: E1203 08:39:27.388740 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.393754 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.393998 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.394012 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.394036 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.394118 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: E1203 08:39:27.411878 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.416952 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.417004 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.417019 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.417040 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.417075 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: E1203 08:39:27.436580 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.443501 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.443558 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.443570 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.443596 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.443609 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: E1203 08:39:27.462919 4573 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56145332-ffe2-4521-8b79-5f3a443b6d75\\\",\\\"systemUUID\\\":\\\"05fa6cec-0bc2-44ec-8da4-d86d857e3ca3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:27Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:27 crc kubenswrapper[4573]: E1203 08:39:27.463118 4573 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.465637 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.465672 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.465682 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.465702 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.465715 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.569114 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.569151 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.569163 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.569181 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.569191 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.671170 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.671243 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.671264 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.671286 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.671303 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.773473 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.773516 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.773525 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.773542 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.773553 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.876844 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.876878 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.876887 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.876908 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.876946 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.979671 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.979764 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.979779 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.979803 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:27 crc kubenswrapper[4573]: I1203 08:39:27.979820 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:27Z","lastTransitionTime":"2025-12-03T08:39:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.030501 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:28 crc kubenswrapper[4573]: E1203 08:39:28.030676 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.030731 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:28 crc kubenswrapper[4573]: E1203 08:39:28.031117 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.083648 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.083701 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.083715 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.083738 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.083755 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.186592 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.186670 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.186696 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.186730 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.186754 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.289074 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.289108 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.289118 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.289133 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.289144 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.392518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.392555 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.392566 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.392582 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.392593 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.495438 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.495481 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.495492 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.495506 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.495517 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.598152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.598196 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.598207 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.598221 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.598232 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.701522 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.701588 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.701645 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.701685 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.701704 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.805164 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.805238 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.805262 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.805295 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.805320 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.907970 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.908036 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.908092 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.908120 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:28 crc kubenswrapper[4573]: I1203 08:39:28.908138 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:28Z","lastTransitionTime":"2025-12-03T08:39:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.010966 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.011030 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.011077 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.011102 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.011121 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.030125 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.030130 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:29 crc kubenswrapper[4573]: E1203 08:39:29.030429 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:29 crc kubenswrapper[4573]: E1203 08:39:29.030299 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.113611 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.113687 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.113702 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.113720 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.113759 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.217332 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.217371 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.217380 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.217398 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.217410 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.320393 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.320480 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.320502 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.320531 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.320549 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.423673 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.423726 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.423739 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.423761 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.423775 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.526318 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.526380 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.526398 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.526426 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.526443 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.628856 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.628911 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.628922 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.628944 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.628960 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.731280 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.731335 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.731349 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.731370 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.731385 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.833718 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.833771 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.833784 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.833802 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.833816 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.936295 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.936353 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.936366 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.936388 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:29 crc kubenswrapper[4573]: I1203 08:39:29.936400 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:29Z","lastTransitionTime":"2025-12-03T08:39:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.029940 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:30 crc kubenswrapper[4573]: E1203 08:39:30.030198 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.030697 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:30 crc kubenswrapper[4573]: E1203 08:39:30.030888 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.039730 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.039789 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.039811 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.039838 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.039862 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.051298 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-ch24s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:39:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T08:39:09Z\\\",\\\"message\\\":\\\"2025-12-03T08:38:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f\\\\n2025-12-03T08:38:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_ab9decdf-cba9-4e29-b851-898a77e4633f to /host/opt/cni/bin/\\\\n2025-12-03T08:38:24Z [verbose] multus-daemon started\\\\n2025-12-03T08:38:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T08:39:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:39:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-multus\"/\"multus-ch24s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.069325 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e17325e8-2a46-4028-b7c5-664eeda96013\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05dc04edf5d1b78774c65964deee9e8a1ebbbf4fc2d28619cfe8bb682cdd35fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xg4ms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.087582 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zxsgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73da931-0de5-4fd1-b8ff-878cc7008bf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://937fc93a7986888d4dbf6b11d9d9f8a57c1c009fbff351c91fa6d794152e69ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-89drg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:19Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zxsgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.107173 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bdff5b8e-11e4-4def-a585-842ecb30f940\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58de0d1d71a5fa581d59e6392bab38079fa755f06bb1805a317ec4afe9abe25b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4d62d2d4e623ef8418da65a2510e8612170cb95f7d9e64ec90c6a6a57417ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8dx9r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-q298n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.125488 4573 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"366494bc-5e98-4a0e-90dd-0eb29f3d975e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T08:38:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b830838fefe3048695aaead38e39afdf274fe2941d2b8b202760f4013b1f41f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53863a744a612591a5e1ddc0ab875d7cf5d86d8baaffa9f09efcede57715a0ab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8bc722a3b0a9f251444556d750fd33f667968dfcb80594ff2b449f5cf515328c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T08:38:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T08:38:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T08:39:30Z is after 2025-08-24T17:21:41Z" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.142764 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.142807 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.142818 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.142836 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.142851 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.225075 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-blvnm" podStartSLOduration=71.225034579 podStartE2EDuration="1m11.225034579s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:30.224391132 +0000 UTC m=+90.792770401" watchObservedRunningTime="2025-12-03 08:39:30.225034579 +0000 UTC m=+90.793413848" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.225199 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=10.225194073 podStartE2EDuration="10.225194073s" podCreationTimestamp="2025-12-03 08:39:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:30.207575635 +0000 UTC m=+90.775954924" watchObservedRunningTime="2025-12-03 08:39:30.225194073 +0000 UTC m=+90.793573342" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.245322 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.245393 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.245407 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.245434 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.245450 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.315766 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-mfcvl" podStartSLOduration=71.315736442 podStartE2EDuration="1m11.315736442s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:30.302099479 +0000 UTC m=+90.870478748" watchObservedRunningTime="2025-12-03 08:39:30.315736442 +0000 UTC m=+90.884115701" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.348438 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.348477 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.348490 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.348509 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.348532 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.371666 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=18.371638549 podStartE2EDuration="18.371638549s" podCreationTimestamp="2025-12-03 08:39:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:30.371056823 +0000 UTC m=+90.939436082" watchObservedRunningTime="2025-12-03 08:39:30.371638549 +0000 UTC m=+90.940017808" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.389864 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.389834343 podStartE2EDuration="1m11.389834343s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:30.389559336 +0000 UTC m=+90.957938595" watchObservedRunningTime="2025-12-03 08:39:30.389834343 +0000 UTC m=+90.958213602" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.406202 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=42.406171338 podStartE2EDuration="42.406171338s" podCreationTimestamp="2025-12-03 08:38:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:30.405447199 +0000 UTC m=+90.973826468" watchObservedRunningTime="2025-12-03 08:39:30.406171338 +0000 UTC m=+90.974550607" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.451113 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.451520 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.451592 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.451659 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.451754 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.554801 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.554842 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.554852 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.554869 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.554879 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.657791 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.657845 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.657862 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.657886 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.657903 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.761584 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.761631 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.761649 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.761674 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.761692 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.864764 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.864851 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.864880 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.864918 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.864941 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.969247 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.969304 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.969321 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.969345 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:30 crc kubenswrapper[4573]: I1203 08:39:30.969363 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:30Z","lastTransitionTime":"2025-12-03T08:39:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.029693 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:31 crc kubenswrapper[4573]: E1203 08:39:31.029870 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.031145 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:31 crc kubenswrapper[4573]: E1203 08:39:31.031249 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.073461 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.073511 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.073528 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.073552 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.073570 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.178252 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.178331 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.178356 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.178387 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.178409 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.281623 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.281676 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.281688 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.281709 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.281770 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.384402 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.384460 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.384477 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.384500 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.384515 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.487983 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.488501 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.488578 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.488656 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.488725 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.591832 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.591901 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.591918 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.591947 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.591968 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.695519 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.695597 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.695620 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.695650 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.695671 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.799331 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.799703 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.799801 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.799873 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.799945 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.912277 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.912345 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.912360 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.912383 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:31 crc kubenswrapper[4573]: I1203 08:39:31.912398 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:31Z","lastTransitionTime":"2025-12-03T08:39:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.016089 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.016155 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.016195 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.016220 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.016234 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.030378 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.030423 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:32 crc kubenswrapper[4573]: E1203 08:39:32.030594 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:32 crc kubenswrapper[4573]: E1203 08:39:32.030720 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.031669 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:39:32 crc kubenswrapper[4573]: E1203 08:39:32.031921 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.069535 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podStartSLOduration=73.069514547 podStartE2EDuration="1m13.069514547s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:32.069160978 +0000 UTC m=+92.637540237" watchObservedRunningTime="2025-12-03 08:39:32.069514547 +0000 UTC m=+92.637893816" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.069682 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-ch24s" podStartSLOduration=73.069676862 podStartE2EDuration="1m13.069676862s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:32.052491724 +0000 UTC m=+92.620870983" watchObservedRunningTime="2025-12-03 08:39:32.069676862 +0000 UTC m=+92.638056131" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.100498 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-zxsgx" podStartSLOduration=73.100471571 podStartE2EDuration="1m13.100471571s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:32.084103935 +0000 UTC m=+92.652483194" watchObservedRunningTime="2025-12-03 08:39:32.100471571 +0000 UTC m=+92.668850830" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.119433 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.119563 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.119623 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.119649 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.119698 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.120536 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-q298n" podStartSLOduration=71.120513484 podStartE2EDuration="1m11.120513484s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:32.101397365 +0000 UTC m=+92.669776644" watchObservedRunningTime="2025-12-03 08:39:32.120513484 +0000 UTC m=+92.688892743" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.170772 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.17073329 podStartE2EDuration="1m9.17073329s" podCreationTimestamp="2025-12-03 08:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:32.121229963 +0000 UTC m=+92.689609232" watchObservedRunningTime="2025-12-03 08:39:32.17073329 +0000 UTC m=+92.739112549" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.223772 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.223824 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.223838 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.223859 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.223874 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.327242 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.327301 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.327315 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.327339 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.327354 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.431215 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.431273 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.431287 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.431313 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.431329 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.534825 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.534898 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.534918 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.534947 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.534967 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.639793 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.639935 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.639965 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.640004 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.640031 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.743610 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.743680 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.743699 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.743729 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.743746 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.846362 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.846438 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.846455 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.846479 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.846496 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.949466 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.949992 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.950163 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.950365 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:32 crc kubenswrapper[4573]: I1203 08:39:32.950520 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:32Z","lastTransitionTime":"2025-12-03T08:39:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.030402 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.030553 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:33 crc kubenswrapper[4573]: E1203 08:39:33.030600 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:33 crc kubenswrapper[4573]: E1203 08:39:33.030787 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.053726 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.053792 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.053802 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.053819 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.053833 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.156902 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.156979 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.156995 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.157021 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.157077 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.260691 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.260745 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.260760 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.260784 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.260801 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.363874 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.363921 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.363932 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.363946 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.363955 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.467921 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.468001 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.468015 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.468034 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.468101 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.570992 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.571039 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.571087 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.571117 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.571141 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.674467 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.674530 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.674541 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.674562 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.674573 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.777503 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.777799 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.777894 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.778001 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.778118 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.882040 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.882142 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.882157 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.882179 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.882194 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.986670 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.986723 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.986750 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.986773 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:33 crc kubenswrapper[4573]: I1203 08:39:33.986783 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:33Z","lastTransitionTime":"2025-12-03T08:39:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.030317 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:34 crc kubenswrapper[4573]: E1203 08:39:34.030522 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.030555 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:34 crc kubenswrapper[4573]: E1203 08:39:34.030765 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.089635 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.089686 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.089728 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.089752 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.089767 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.192280 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.192356 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.192368 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.192389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.192403 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.295153 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.295203 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.295214 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.295230 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.295242 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.398520 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.398589 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.398601 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.398620 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.398672 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.501149 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.501185 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.501194 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.501209 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.501220 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.603750 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.603852 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.603890 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.603926 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.603952 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.707328 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.707406 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.707418 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.707432 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.707469 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.809807 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.809858 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.809870 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.809888 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.809899 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.913070 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.913361 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.913492 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.913596 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:34 crc kubenswrapper[4573]: I1203 08:39:34.913688 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:34Z","lastTransitionTime":"2025-12-03T08:39:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.016742 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.017093 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.017250 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.017382 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.017477 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.029656 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:35 crc kubenswrapper[4573]: E1203 08:39:35.030440 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.030756 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:35 crc kubenswrapper[4573]: E1203 08:39:35.030892 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.133473 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.133512 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.133522 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.133540 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.133552 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.235550 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.235597 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.235612 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.235634 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.235646 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.337905 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.338227 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.338249 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.338272 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.338294 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.440527 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.440595 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.440620 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.440649 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.440675 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.543784 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.543855 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.543869 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.543919 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.543936 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.646760 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.646797 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.646808 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.646823 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.646837 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.749987 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.750072 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.750091 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.750115 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.750130 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.853083 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.853147 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.853166 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.853188 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.853200 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.956031 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.956137 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.956152 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.956177 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:35 crc kubenswrapper[4573]: I1203 08:39:35.956190 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:35Z","lastTransitionTime":"2025-12-03T08:39:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.029745 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.029779 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:36 crc kubenswrapper[4573]: E1203 08:39:36.030026 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:36 crc kubenswrapper[4573]: E1203 08:39:36.030225 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.059844 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.059887 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.059896 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.059910 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.059923 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.163019 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.163098 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.163115 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.163134 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.163148 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.266483 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.266518 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.266526 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.266543 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.266552 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.368886 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.369367 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.369475 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.369591 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.369685 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.473468 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.473519 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.473532 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.473552 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.473565 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.575709 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.575742 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.575750 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.575763 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.575772 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.678706 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.678749 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.678761 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.678778 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.678792 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.782636 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.782687 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.782699 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.782715 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.782730 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.886447 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.886526 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.886551 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.886581 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.886604 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.990340 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.990398 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.990411 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.990433 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:36 crc kubenswrapper[4573]: I1203 08:39:36.990450 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:36Z","lastTransitionTime":"2025-12-03T08:39:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.030019 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.030114 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:37 crc kubenswrapper[4573]: E1203 08:39:37.030294 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:37 crc kubenswrapper[4573]: E1203 08:39:37.030422 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.093540 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.093602 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.093615 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.093636 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.093656 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:37Z","lastTransitionTime":"2025-12-03T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.197227 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.197639 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.197780 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.197893 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.197981 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:37Z","lastTransitionTime":"2025-12-03T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.302374 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.302424 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.302435 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.302452 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.302464 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:37Z","lastTransitionTime":"2025-12-03T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.405303 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.405372 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.405389 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.405416 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.405433 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:37Z","lastTransitionTime":"2025-12-03T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.508271 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.508330 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.508348 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.508374 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.508393 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:37Z","lastTransitionTime":"2025-12-03T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.611337 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.611400 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.611415 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.611433 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.611448 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:37Z","lastTransitionTime":"2025-12-03T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.613867 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.613941 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.613963 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.613988 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.614007 4573 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T08:39:37Z","lastTransitionTime":"2025-12-03T08:39:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.668271 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5"] Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.668904 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.671127 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.673247 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.673452 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.673765 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.764477 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.764616 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.764651 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.764695 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.764774 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.866021 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.866088 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.866107 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.866125 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.866139 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.866438 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.867196 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.867241 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.876199 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.892811 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7b1ac5bc-26bf-42d6-aa8b-0894925911c4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j99r5\" (UID: \"7b1ac5bc-26bf-42d6-aa8b-0894925911c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:37 crc kubenswrapper[4573]: I1203 08:39:37.993343 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" Dec 03 08:39:38 crc kubenswrapper[4573]: W1203 08:39:38.018238 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b1ac5bc_26bf_42d6_aa8b_0894925911c4.slice/crio-626c5a3793617060e5af800f9816d5321615abcec9621a50fff3160a1c070cae WatchSource:0}: Error finding container 626c5a3793617060e5af800f9816d5321615abcec9621a50fff3160a1c070cae: Status 404 returned error can't find the container with id 626c5a3793617060e5af800f9816d5321615abcec9621a50fff3160a1c070cae Dec 03 08:39:38 crc kubenswrapper[4573]: I1203 08:39:38.030825 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:38 crc kubenswrapper[4573]: E1203 08:39:38.031106 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:38 crc kubenswrapper[4573]: I1203 08:39:38.031482 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:38 crc kubenswrapper[4573]: E1203 08:39:38.031922 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:38 crc kubenswrapper[4573]: I1203 08:39:38.575629 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:38 crc kubenswrapper[4573]: E1203 08:39:38.575770 4573 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:39:38 crc kubenswrapper[4573]: E1203 08:39:38.575838 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs podName:bb7f6603-0f41-4e0f-b77e-47dbcc676aaf nodeName:}" failed. No retries permitted until 2025-12-03 08:40:42.575819861 +0000 UTC m=+163.144199130 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs") pod "network-metrics-daemon-kcphv" (UID: "bb7f6603-0f41-4e0f-b77e-47dbcc676aaf") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 08:39:38 crc kubenswrapper[4573]: I1203 08:39:38.937915 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" event={"ID":"7b1ac5bc-26bf-42d6-aa8b-0894925911c4","Type":"ContainerStarted","Data":"2a404d6f0cbb84159c2d949fde11b6fd723bef670386ffb95480750c43b22c4e"} Dec 03 08:39:38 crc kubenswrapper[4573]: I1203 08:39:38.937993 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" event={"ID":"7b1ac5bc-26bf-42d6-aa8b-0894925911c4","Type":"ContainerStarted","Data":"626c5a3793617060e5af800f9816d5321615abcec9621a50fff3160a1c070cae"} Dec 03 08:39:39 crc kubenswrapper[4573]: I1203 08:39:39.029830 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:39 crc kubenswrapper[4573]: I1203 08:39:39.029842 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:39 crc kubenswrapper[4573]: E1203 08:39:39.030005 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:39 crc kubenswrapper[4573]: E1203 08:39:39.030241 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:40 crc kubenswrapper[4573]: I1203 08:39:40.030315 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:40 crc kubenswrapper[4573]: I1203 08:39:40.030315 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:40 crc kubenswrapper[4573]: E1203 08:39:40.031580 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:40 crc kubenswrapper[4573]: E1203 08:39:40.031730 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:41 crc kubenswrapper[4573]: I1203 08:39:41.030020 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:41 crc kubenswrapper[4573]: I1203 08:39:41.030211 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:41 crc kubenswrapper[4573]: E1203 08:39:41.030348 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:41 crc kubenswrapper[4573]: E1203 08:39:41.030756 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:42 crc kubenswrapper[4573]: I1203 08:39:42.030400 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:42 crc kubenswrapper[4573]: I1203 08:39:42.030400 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:42 crc kubenswrapper[4573]: E1203 08:39:42.030689 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:42 crc kubenswrapper[4573]: E1203 08:39:42.030744 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:43 crc kubenswrapper[4573]: I1203 08:39:43.030335 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:43 crc kubenswrapper[4573]: I1203 08:39:43.030451 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:43 crc kubenswrapper[4573]: E1203 08:39:43.030537 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:43 crc kubenswrapper[4573]: E1203 08:39:43.030687 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:44 crc kubenswrapper[4573]: I1203 08:39:44.030836 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:44 crc kubenswrapper[4573]: E1203 08:39:44.031105 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:44 crc kubenswrapper[4573]: I1203 08:39:44.030137 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:44 crc kubenswrapper[4573]: E1203 08:39:44.031625 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:44 crc kubenswrapper[4573]: I1203 08:39:44.039708 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:39:44 crc kubenswrapper[4573]: E1203 08:39:44.041801 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:39:45 crc kubenswrapper[4573]: I1203 08:39:45.029443 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:45 crc kubenswrapper[4573]: E1203 08:39:45.029592 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:45 crc kubenswrapper[4573]: I1203 08:39:45.029452 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:45 crc kubenswrapper[4573]: E1203 08:39:45.029669 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:46 crc kubenswrapper[4573]: I1203 08:39:46.030504 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:46 crc kubenswrapper[4573]: E1203 08:39:46.030781 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:46 crc kubenswrapper[4573]: I1203 08:39:46.031285 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:46 crc kubenswrapper[4573]: E1203 08:39:46.031510 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:47 crc kubenswrapper[4573]: I1203 08:39:47.030338 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:47 crc kubenswrapper[4573]: I1203 08:39:47.030338 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:47 crc kubenswrapper[4573]: E1203 08:39:47.030523 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:47 crc kubenswrapper[4573]: E1203 08:39:47.030626 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:48 crc kubenswrapper[4573]: I1203 08:39:48.030119 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:48 crc kubenswrapper[4573]: I1203 08:39:48.030158 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:48 crc kubenswrapper[4573]: E1203 08:39:48.030316 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:48 crc kubenswrapper[4573]: E1203 08:39:48.031014 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:49 crc kubenswrapper[4573]: I1203 08:39:49.030283 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:49 crc kubenswrapper[4573]: I1203 08:39:49.030336 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:49 crc kubenswrapper[4573]: E1203 08:39:49.030492 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:49 crc kubenswrapper[4573]: E1203 08:39:49.030530 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:50 crc kubenswrapper[4573]: I1203 08:39:50.029786 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:50 crc kubenswrapper[4573]: I1203 08:39:50.029786 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:50 crc kubenswrapper[4573]: E1203 08:39:50.031091 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:50 crc kubenswrapper[4573]: E1203 08:39:50.031216 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:51 crc kubenswrapper[4573]: I1203 08:39:51.029854 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:51 crc kubenswrapper[4573]: I1203 08:39:51.029937 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:51 crc kubenswrapper[4573]: E1203 08:39:51.030042 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:51 crc kubenswrapper[4573]: E1203 08:39:51.030220 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:52 crc kubenswrapper[4573]: I1203 08:39:52.030476 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:52 crc kubenswrapper[4573]: I1203 08:39:52.030555 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:52 crc kubenswrapper[4573]: E1203 08:39:52.030739 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:52 crc kubenswrapper[4573]: E1203 08:39:52.030906 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:53 crc kubenswrapper[4573]: I1203 08:39:53.030025 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:53 crc kubenswrapper[4573]: I1203 08:39:53.030016 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:53 crc kubenswrapper[4573]: E1203 08:39:53.030248 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:53 crc kubenswrapper[4573]: E1203 08:39:53.030550 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:54 crc kubenswrapper[4573]: I1203 08:39:54.029411 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:54 crc kubenswrapper[4573]: E1203 08:39:54.029693 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:54 crc kubenswrapper[4573]: I1203 08:39:54.029746 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:54 crc kubenswrapper[4573]: E1203 08:39:54.030188 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:55 crc kubenswrapper[4573]: I1203 08:39:55.029840 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:55 crc kubenswrapper[4573]: I1203 08:39:55.029870 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:55 crc kubenswrapper[4573]: E1203 08:39:55.030016 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:55 crc kubenswrapper[4573]: E1203 08:39:55.030537 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:55 crc kubenswrapper[4573]: I1203 08:39:55.030821 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:39:55 crc kubenswrapper[4573]: E1203 08:39:55.031004 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x466z_openshift-ovn-kubernetes(48186949-d112-4fb7-aa60-026d0176c73e)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.002003 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/1.log" Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.003604 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/0.log" Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.003687 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c98ef96-0e91-49e3-88d5-c95dd6f52c5d" containerID="35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0" exitCode=1 Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.003752 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ch24s" event={"ID":"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d","Type":"ContainerDied","Data":"35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0"} Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.003858 4573 scope.go:117] "RemoveContainer" containerID="296a87ae21056780713b53c9cd6c1f73573ba7f79fa02189631d79bef8dcdec0" Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.004444 4573 scope.go:117] "RemoveContainer" containerID="35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0" Dec 03 08:39:56 crc kubenswrapper[4573]: E1203 08:39:56.006659 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-ch24s_openshift-multus(8c98ef96-0e91-49e3-88d5-c95dd6f52c5d)\"" pod="openshift-multus/multus-ch24s" podUID="8c98ef96-0e91-49e3-88d5-c95dd6f52c5d" Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.033026 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j99r5" podStartSLOduration=97.032961559 podStartE2EDuration="1m37.032961559s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:39:38.963396133 +0000 UTC m=+99.531775472" watchObservedRunningTime="2025-12-03 08:39:56.032961559 +0000 UTC m=+116.601340818" Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.034686 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:56 crc kubenswrapper[4573]: E1203 08:39:56.034931 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:56 crc kubenswrapper[4573]: I1203 08:39:56.035247 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:56 crc kubenswrapper[4573]: E1203 08:39:56.035314 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:57 crc kubenswrapper[4573]: I1203 08:39:57.010166 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/1.log" Dec 03 08:39:57 crc kubenswrapper[4573]: I1203 08:39:57.029791 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:57 crc kubenswrapper[4573]: I1203 08:39:57.030291 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:57 crc kubenswrapper[4573]: E1203 08:39:57.030662 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:57 crc kubenswrapper[4573]: E1203 08:39:57.030826 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:58 crc kubenswrapper[4573]: I1203 08:39:58.030134 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:39:58 crc kubenswrapper[4573]: I1203 08:39:58.030177 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:39:58 crc kubenswrapper[4573]: E1203 08:39:58.030540 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:39:58 crc kubenswrapper[4573]: E1203 08:39:58.030454 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:39:59 crc kubenswrapper[4573]: I1203 08:39:59.030898 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:39:59 crc kubenswrapper[4573]: I1203 08:39:59.030915 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:39:59 crc kubenswrapper[4573]: E1203 08:39:59.031085 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:39:59 crc kubenswrapper[4573]: E1203 08:39:59.031193 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:39:59 crc kubenswrapper[4573]: E1203 08:39:59.949998 4573 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 08:40:00 crc kubenswrapper[4573]: I1203 08:40:00.031720 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:00 crc kubenswrapper[4573]: E1203 08:40:00.031866 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:40:00 crc kubenswrapper[4573]: I1203 08:40:00.032255 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:00 crc kubenswrapper[4573]: E1203 08:40:00.032387 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:00 crc kubenswrapper[4573]: E1203 08:40:00.139836 4573 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 08:40:01 crc kubenswrapper[4573]: I1203 08:40:01.030015 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:01 crc kubenswrapper[4573]: I1203 08:40:01.030075 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:01 crc kubenswrapper[4573]: E1203 08:40:01.030608 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:40:01 crc kubenswrapper[4573]: E1203 08:40:01.030636 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:40:02 crc kubenswrapper[4573]: I1203 08:40:02.029677 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:02 crc kubenswrapper[4573]: I1203 08:40:02.029714 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:02 crc kubenswrapper[4573]: E1203 08:40:02.029916 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:02 crc kubenswrapper[4573]: E1203 08:40:02.030194 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:40:03 crc kubenswrapper[4573]: I1203 08:40:03.029819 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:03 crc kubenswrapper[4573]: I1203 08:40:03.029888 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:03 crc kubenswrapper[4573]: E1203 08:40:03.030997 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:40:03 crc kubenswrapper[4573]: E1203 08:40:03.031319 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:40:04 crc kubenswrapper[4573]: I1203 08:40:04.030001 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:04 crc kubenswrapper[4573]: I1203 08:40:04.030163 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:04 crc kubenswrapper[4573]: E1203 08:40:04.030303 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:04 crc kubenswrapper[4573]: E1203 08:40:04.030483 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:40:05 crc kubenswrapper[4573]: I1203 08:40:05.029465 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:05 crc kubenswrapper[4573]: I1203 08:40:05.029489 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:05 crc kubenswrapper[4573]: E1203 08:40:05.029706 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:40:05 crc kubenswrapper[4573]: E1203 08:40:05.029840 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:40:05 crc kubenswrapper[4573]: E1203 08:40:05.141632 4573 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 08:40:06 crc kubenswrapper[4573]: I1203 08:40:06.029722 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:06 crc kubenswrapper[4573]: E1203 08:40:06.029931 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:40:06 crc kubenswrapper[4573]: I1203 08:40:06.029744 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:06 crc kubenswrapper[4573]: E1203 08:40:06.030232 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:07 crc kubenswrapper[4573]: I1203 08:40:07.030278 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:07 crc kubenswrapper[4573]: I1203 08:40:07.030454 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:07 crc kubenswrapper[4573]: I1203 08:40:07.030579 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:40:07 crc kubenswrapper[4573]: E1203 08:40:07.031613 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:40:07 crc kubenswrapper[4573]: E1203 08:40:07.031673 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:40:08 crc kubenswrapper[4573]: I1203 08:40:08.032386 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:08 crc kubenswrapper[4573]: E1203 08:40:08.032975 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:08 crc kubenswrapper[4573]: I1203 08:40:08.033115 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:08 crc kubenswrapper[4573]: E1203 08:40:08.033186 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:40:08 crc kubenswrapper[4573]: I1203 08:40:08.056516 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/3.log" Dec 03 08:40:08 crc kubenswrapper[4573]: I1203 08:40:08.059179 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerStarted","Data":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} Dec 03 08:40:08 crc kubenswrapper[4573]: I1203 08:40:08.060664 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:40:08 crc kubenswrapper[4573]: I1203 08:40:08.106353 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podStartSLOduration=109.106325493 podStartE2EDuration="1m49.106325493s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:08.104015422 +0000 UTC m=+128.672394691" watchObservedRunningTime="2025-12-03 08:40:08.106325493 +0000 UTC m=+128.674704752" Dec 03 08:40:09 crc kubenswrapper[4573]: I1203 08:40:09.049986 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:09 crc kubenswrapper[4573]: I1203 08:40:09.050161 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:09 crc kubenswrapper[4573]: I1203 08:40:09.050557 4573 scope.go:117] "RemoveContainer" containerID="35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0" Dec 03 08:40:09 crc kubenswrapper[4573]: E1203 08:40:09.050940 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:40:09 crc kubenswrapper[4573]: E1203 08:40:09.050981 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:40:09 crc kubenswrapper[4573]: I1203 08:40:09.106781 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kcphv"] Dec 03 08:40:09 crc kubenswrapper[4573]: I1203 08:40:09.106927 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:09 crc kubenswrapper[4573]: E1203 08:40:09.107035 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:10 crc kubenswrapper[4573]: I1203 08:40:10.031210 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:10 crc kubenswrapper[4573]: E1203 08:40:10.034271 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:40:10 crc kubenswrapper[4573]: I1203 08:40:10.071913 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/1.log" Dec 03 08:40:10 crc kubenswrapper[4573]: I1203 08:40:10.071985 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ch24s" event={"ID":"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d","Type":"ContainerStarted","Data":"dcf9f6f434fdad7cbbf3d2a5bda9998bad92078f8bb45c93cdb1b3fc856089c1"} Dec 03 08:40:10 crc kubenswrapper[4573]: E1203 08:40:10.142490 4573 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 08:40:11 crc kubenswrapper[4573]: I1203 08:40:11.030225 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:11 crc kubenswrapper[4573]: E1203 08:40:11.030368 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:40:11 crc kubenswrapper[4573]: I1203 08:40:11.030566 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:11 crc kubenswrapper[4573]: E1203 08:40:11.030629 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:40:11 crc kubenswrapper[4573]: I1203 08:40:11.030762 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:11 crc kubenswrapper[4573]: E1203 08:40:11.030822 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:12 crc kubenswrapper[4573]: I1203 08:40:12.029793 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:12 crc kubenswrapper[4573]: E1203 08:40:12.029963 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:40:13 crc kubenswrapper[4573]: I1203 08:40:13.029533 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:13 crc kubenswrapper[4573]: I1203 08:40:13.029539 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:13 crc kubenswrapper[4573]: E1203 08:40:13.029681 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:40:13 crc kubenswrapper[4573]: I1203 08:40:13.029533 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:13 crc kubenswrapper[4573]: E1203 08:40:13.029771 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:13 crc kubenswrapper[4573]: E1203 08:40:13.029811 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:40:14 crc kubenswrapper[4573]: I1203 08:40:14.029593 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:14 crc kubenswrapper[4573]: E1203 08:40:14.029781 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 08:40:15 crc kubenswrapper[4573]: I1203 08:40:15.030002 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:15 crc kubenswrapper[4573]: I1203 08:40:15.030098 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:15 crc kubenswrapper[4573]: E1203 08:40:15.030200 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kcphv" podUID="bb7f6603-0f41-4e0f-b77e-47dbcc676aaf" Dec 03 08:40:15 crc kubenswrapper[4573]: E1203 08:40:15.030292 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 08:40:15 crc kubenswrapper[4573]: I1203 08:40:15.031384 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:15 crc kubenswrapper[4573]: E1203 08:40:15.031658 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 08:40:16 crc kubenswrapper[4573]: I1203 08:40:16.029765 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:16 crc kubenswrapper[4573]: I1203 08:40:16.036593 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 08:40:16 crc kubenswrapper[4573]: I1203 08:40:16.038772 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 08:40:17 crc kubenswrapper[4573]: I1203 08:40:17.029409 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:17 crc kubenswrapper[4573]: I1203 08:40:17.029457 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:17 crc kubenswrapper[4573]: I1203 08:40:17.029527 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:17 crc kubenswrapper[4573]: I1203 08:40:17.031676 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 08:40:17 crc kubenswrapper[4573]: I1203 08:40:17.032538 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 08:40:17 crc kubenswrapper[4573]: I1203 08:40:17.032619 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 08:40:17 crc kubenswrapper[4573]: I1203 08:40:17.033839 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.330855 4573 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.400496 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.401010 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.402976 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6wlqk"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.403457 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.404091 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.404531 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.405005 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.405685 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.406114 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.406554 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8qfqs"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.406553 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.406635 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.407968 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.412218 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.413477 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2w8js"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.414059 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.414609 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.419381 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.419510 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.419563 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.419563 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.419680 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.419924 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.420224 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.420872 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vphhd"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.421393 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.421695 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fg658"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.422316 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.422358 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.422322 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.422524 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.423069 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.423073 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.423140 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.423438 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.423925 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.424617 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.424902 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.425297 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.425856 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.426743 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-dpwlm"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.427008 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.435329 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.439274 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.439374 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453414 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-audit\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453460 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2de8f09-187b-4b8f-9799-321d14745677-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453506 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5279cafe-fc2e-4037-8841-ac430e0903b6-audit-dir\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453528 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz9pv\" (UniqueName: \"kubernetes.io/projected/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-kube-api-access-hz9pv\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453549 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-config\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453576 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1783a29-c599-40fe-a075-32732bb082db-audit-dir\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453614 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-etcd-client\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453638 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bpl4\" (UniqueName: \"kubernetes.io/projected/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-kube-api-access-2bpl4\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453657 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b5ee846-009f-4feb-a61b-a413aee18c4d-serving-cert\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453679 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh52v\" (UniqueName: \"kubernetes.io/projected/284d9db2-04e5-4c39-b4cf-6f001de356c4-kube-api-access-bh52v\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453722 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqgrt\" (UniqueName: \"kubernetes.io/projected/9b31f7fa-c98a-45e3-934e-94e70d254fca-kube-api-access-vqgrt\") pod \"cluster-samples-operator-665b6dd947-sjs5t\" (UID: \"9b31f7fa-c98a-45e3-934e-94e70d254fca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453750 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-ca\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453769 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-config\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453790 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7807daee-9b44-4ebd-aab9-afaf91fb9746-auth-proxy-config\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453808 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7807daee-9b44-4ebd-aab9-afaf91fb9746-machine-approver-tls\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453839 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-serving-cert\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453871 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8b5ee846-009f-4feb-a61b-a413aee18c4d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453894 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/284d9db2-04e5-4c39-b4cf-6f001de356c4-images\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453913 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59pwc\" (UniqueName: \"kubernetes.io/projected/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-kube-api-access-59pwc\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453933 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-service-ca\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453969 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-serving-cert\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.453987 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-trusted-ca\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454007 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8p4l\" (UniqueName: \"kubernetes.io/projected/ee739812-3e92-42ff-a39d-fcea79a3164c-kube-api-access-n8p4l\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454026 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsgx9\" (UniqueName: \"kubernetes.io/projected/7807daee-9b44-4ebd-aab9-afaf91fb9746-kube-api-access-zsgx9\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454093 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-oauth-config\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454115 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/284d9db2-04e5-4c39-b4cf-6f001de356c4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454136 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2de8f09-187b-4b8f-9799-321d14745677-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454187 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-serving-cert\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454237 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-audit-policies\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454273 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-client\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454309 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-config\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454323 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7807daee-9b44-4ebd-aab9-afaf91fb9746-config\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454341 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljd6p\" (UniqueName: \"kubernetes.io/projected/5279cafe-fc2e-4037-8841-ac430e0903b6-kube-api-access-ljd6p\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454394 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5279cafe-fc2e-4037-8841-ac430e0903b6-node-pullsecrets\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454432 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-trusted-ca-bundle\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454468 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-oauth-serving-cert\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454507 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-etcd-client\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454532 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454555 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-encryption-config\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454591 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67d2h\" (UniqueName: \"kubernetes.io/projected/cc4dd1dc-155a-4278-8382-cbe5364f80eb-kube-api-access-67d2h\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454614 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b31f7fa-c98a-45e3-934e-94e70d254fca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sjs5t\" (UID: \"9b31f7fa-c98a-45e3-934e-94e70d254fca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454634 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-client-ca\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454750 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn2q5\" (UniqueName: \"kubernetes.io/projected/8b5ee846-009f-4feb-a61b-a413aee18c4d-kube-api-access-xn2q5\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454782 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454815 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-service-ca\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454836 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/284d9db2-04e5-4c39-b4cf-6f001de356c4-config\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454858 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee739812-3e92-42ff-a39d-fcea79a3164c-serving-cert\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454929 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-client-ca\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454951 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-serving-cert\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454975 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-config\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.454996 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-image-import-ca\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455028 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crxbp\" (UniqueName: \"kubernetes.io/projected/f1783a29-c599-40fe-a075-32732bb082db-kube-api-access-crxbp\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455051 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22grw\" (UniqueName: \"kubernetes.io/projected/f2de8f09-187b-4b8f-9799-321d14745677-kube-api-access-22grw\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455120 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-config\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455163 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-etcd-serving-ca\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455179 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455193 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455210 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-config\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455247 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc4dd1dc-155a-4278-8382-cbe5364f80eb-serving-cert\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455266 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-encryption-config\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.455282 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-serving-cert\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.471937 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472209 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472368 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472426 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472562 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472630 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472672 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472636 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472748 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472902 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.472916 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.473228 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.493225 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.493322 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.493508 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.494773 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.495037 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-b92rg"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.495316 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.495571 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.495736 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.495905 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.496194 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.496259 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.496315 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.496377 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.496417 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.496570 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.496830 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.496934 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.497034 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.497140 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.497222 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.503426 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.513439 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.513452 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.513452 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.513544 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.513901 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.514092 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.514278 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.514413 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.514434 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.514514 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.514606 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.514889 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.515001 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.515101 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.515177 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.516565 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vzjzb"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.517216 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.521852 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.521872 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.521937 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.521852 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.522011 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.522227 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.522270 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.523241 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.524503 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.524884 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.525094 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.526979 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h7gd4"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.527349 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.527385 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.527459 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.527537 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.527577 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.528247 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.528422 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.528602 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.528875 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.529056 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.530351 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m7kqr"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.530778 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.532470 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.536339 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.536474 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.536691 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.538978 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.539418 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.541347 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.542158 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.546840 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.546853 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.554482 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575757 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2de8f09-187b-4b8f-9799-321d14745677-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575785 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575798 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-oauth-config\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575826 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/284d9db2-04e5-4c39-b4cf-6f001de356c4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575852 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-serving-cert\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575874 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-audit-policies\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575896 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-client\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575916 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-config\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575948 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7807daee-9b44-4ebd-aab9-afaf91fb9746-config\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575972 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljd6p\" (UniqueName: \"kubernetes.io/projected/5279cafe-fc2e-4037-8841-ac430e0903b6-kube-api-access-ljd6p\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.575992 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5279cafe-fc2e-4037-8841-ac430e0903b6-node-pullsecrets\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576011 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-oauth-serving-cert\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576029 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-etcd-client\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576054 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576095 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-encryption-config\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576121 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fef5501f-3c10-40c0-8f37-0383d15eb68f-metrics-tls\") pod \"dns-operator-744455d44c-vzjzb\" (UID: \"fef5501f-3c10-40c0-8f37-0383d15eb68f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576145 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-trusted-ca-bundle\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576167 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-client-ca\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576189 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67d2h\" (UniqueName: \"kubernetes.io/projected/cc4dd1dc-155a-4278-8382-cbe5364f80eb-kube-api-access-67d2h\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576210 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b31f7fa-c98a-45e3-934e-94e70d254fca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sjs5t\" (UID: \"9b31f7fa-c98a-45e3-934e-94e70d254fca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576240 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn2q5\" (UniqueName: \"kubernetes.io/projected/8b5ee846-009f-4feb-a61b-a413aee18c4d-kube-api-access-xn2q5\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576259 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576279 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/284d9db2-04e5-4c39-b4cf-6f001de356c4-config\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576298 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-service-ca\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576317 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee739812-3e92-42ff-a39d-fcea79a3164c-serving-cert\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576336 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-client-ca\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576362 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-serving-cert\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576380 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-config\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576399 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-image-import-ca\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576418 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crxbp\" (UniqueName: \"kubernetes.io/projected/f1783a29-c599-40fe-a075-32732bb082db-kube-api-access-crxbp\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576435 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22grw\" (UniqueName: \"kubernetes.io/projected/f2de8f09-187b-4b8f-9799-321d14745677-kube-api-access-22grw\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576456 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-config\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576476 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-etcd-serving-ca\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576495 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576512 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576534 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfc8w\" (UniqueName: \"kubernetes.io/projected/fef5501f-3c10-40c0-8f37-0383d15eb68f-kube-api-access-hfc8w\") pod \"dns-operator-744455d44c-vzjzb\" (UID: \"fef5501f-3c10-40c0-8f37-0383d15eb68f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576555 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-config\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576575 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-serving-cert\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576593 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc4dd1dc-155a-4278-8382-cbe5364f80eb-serving-cert\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576610 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-encryption-config\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576639 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-audit\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576661 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2de8f09-187b-4b8f-9799-321d14745677-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576684 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5279cafe-fc2e-4037-8841-ac430e0903b6-audit-dir\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576705 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz9pv\" (UniqueName: \"kubernetes.io/projected/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-kube-api-access-hz9pv\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576734 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-config\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576752 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1783a29-c599-40fe-a075-32732bb082db-audit-dir\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576770 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-etcd-client\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576788 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bpl4\" (UniqueName: \"kubernetes.io/projected/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-kube-api-access-2bpl4\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576810 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqgrt\" (UniqueName: \"kubernetes.io/projected/9b31f7fa-c98a-45e3-934e-94e70d254fca-kube-api-access-vqgrt\") pod \"cluster-samples-operator-665b6dd947-sjs5t\" (UID: \"9b31f7fa-c98a-45e3-934e-94e70d254fca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576831 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b5ee846-009f-4feb-a61b-a413aee18c4d-serving-cert\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576851 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh52v\" (UniqueName: \"kubernetes.io/projected/284d9db2-04e5-4c39-b4cf-6f001de356c4-kube-api-access-bh52v\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576871 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-ca\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576891 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.577573 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7807daee-9b44-4ebd-aab9-afaf91fb9746-auth-proxy-config\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.577623 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.576893 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7807daee-9b44-4ebd-aab9-afaf91fb9746-auth-proxy-config\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.577872 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7807daee-9b44-4ebd-aab9-afaf91fb9746-machine-approver-tls\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.577902 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-config\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.577925 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-serving-cert\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.577951 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8b5ee846-009f-4feb-a61b-a413aee18c4d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.577975 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/284d9db2-04e5-4c39-b4cf-6f001de356c4-images\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.577995 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59pwc\" (UniqueName: \"kubernetes.io/projected/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-kube-api-access-59pwc\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.578016 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-service-ca\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.578039 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-serving-cert\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.578095 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-trusted-ca\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.578118 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8p4l\" (UniqueName: \"kubernetes.io/projected/ee739812-3e92-42ff-a39d-fcea79a3164c-kube-api-access-n8p4l\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.578140 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsgx9\" (UniqueName: \"kubernetes.io/projected/7807daee-9b44-4ebd-aab9-afaf91fb9746-kube-api-access-zsgx9\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.578656 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.580203 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-config\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.580984 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.593423 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.597914 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.598372 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2ktws"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.598764 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.603372 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-audit-policies\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.603505 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-service-ca\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.611797 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-config\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.612335 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7807daee-9b44-4ebd-aab9-afaf91fb9746-config\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.612437 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5279cafe-fc2e-4037-8841-ac430e0903b6-node-pullsecrets\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.613094 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-oauth-serving-cert\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.614037 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-config\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.614464 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-etcd-serving-ca\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.615354 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.620568 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.622513 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5279cafe-fc2e-4037-8841-ac430e0903b6-audit-dir\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.623007 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.623199 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-config\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.623249 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f1783a29-c599-40fe-a075-32732bb082db-audit-dir\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.623868 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.631040 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-client-ca\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.631444 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.631838 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-8tpmr"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.632167 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.633425 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.634445 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.634630 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.638312 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.638653 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.639504 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.640002 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-etcd-client\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.640367 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.640505 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-audit\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.640636 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f1783a29-c599-40fe-a075-32732bb082db-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.640790 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.640964 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.641199 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.641508 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.642075 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-service-ca\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.642934 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-client-ca\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.644143 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-config\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.645008 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5279cafe-fc2e-4037-8841-ac430e0903b6-image-import-ca\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.645664 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-config\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.647480 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f2de8f09-187b-4b8f-9799-321d14745677-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.647745 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-ca\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.648944 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.649172 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.649519 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-encryption-config\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.649706 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc4dd1dc-155a-4278-8382-cbe5364f80eb-serving-cert\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.649708 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pxklw"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.649794 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.650213 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/284d9db2-04e5-4c39-b4cf-6f001de356c4-config\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.650216 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b5ee846-009f-4feb-a61b-a413aee18c4d-serving-cert\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.650448 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.650555 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1783a29-c599-40fe-a075-32732bb082db-serving-cert\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.650992 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-serving-cert\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.651227 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.693597 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8b5ee846-009f-4feb-a61b-a413aee18c4d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.694038 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2de8f09-187b-4b8f-9799-321d14745677-config\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.694724 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/284d9db2-04e5-4c39-b4cf-6f001de356c4-images\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.694814 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-etcd-client\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.695134 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.695414 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b31f7fa-c98a-45e3-934e-94e70d254fca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sjs5t\" (UID: \"9b31f7fa-c98a-45e3-934e-94e70d254fca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.696633 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.697122 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.698176 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fef5501f-3c10-40c0-8f37-0383d15eb68f-metrics-tls\") pod \"dns-operator-744455d44c-vzjzb\" (UID: \"fef5501f-3c10-40c0-8f37-0383d15eb68f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.698354 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfc8w\" (UniqueName: \"kubernetes.io/projected/fef5501f-3c10-40c0-8f37-0383d15eb68f-kube-api-access-hfc8w\") pod \"dns-operator-744455d44c-vzjzb\" (UID: \"fef5501f-3c10-40c0-8f37-0383d15eb68f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.699018 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.701255 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-serving-cert\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.701290 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cc4dd1dc-155a-4278-8382-cbe5364f80eb-etcd-client\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.701908 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/284d9db2-04e5-4c39-b4cf-6f001de356c4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.702024 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.702215 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee739812-3e92-42ff-a39d-fcea79a3164c-serving-cert\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.702281 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k2b4r"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.702533 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-encryption-config\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.702634 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-serving-cert\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.703041 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7807daee-9b44-4ebd-aab9-afaf91fb9746-machine-approver-tls\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.703472 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-oauth-config\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.703753 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5279cafe-fc2e-4037-8841-ac430e0903b6-serving-cert\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.704210 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.704544 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.704645 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.709315 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.706780 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.711345 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-trusted-ca-bundle\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.713429 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6wlqk"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.713623 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.718611 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.710589 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fef5501f-3c10-40c0-8f37-0383d15eb68f-metrics-tls\") pod \"dns-operator-744455d44c-vzjzb\" (UID: \"fef5501f-3c10-40c0-8f37-0383d15eb68f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.729508 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.736084 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.736291 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.737725 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.737871 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.738515 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.738550 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.742974 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.744063 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.744380 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.744518 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.744998 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jdvbk"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.745278 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.745317 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.745550 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.745657 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.746760 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.746894 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2w8js"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.749590 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.750690 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-trusted-ca\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.758237 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vphhd"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.761990 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8qfqs"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.766326 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-smk4x"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.767616 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.768737 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.770129 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-bjl65"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.770679 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bjl65" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.771358 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pxklw"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.773727 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.775275 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.776760 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.778087 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsgx9\" (UniqueName: \"kubernetes.io/projected/7807daee-9b44-4ebd-aab9-afaf91fb9746-kube-api-access-zsgx9\") pod \"machine-approver-56656f9798-5gf25\" (UID: \"7807daee-9b44-4ebd-aab9-afaf91fb9746\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.778238 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b92rg"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.780119 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h7gd4"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.782691 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.784019 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.785173 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-4twfh"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.786105 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.787430 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-92rk6"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.791406 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.791444 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.791524 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.791981 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vzjzb"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.793236 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2ktws"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.794421 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.795531 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m7kqr"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.797232 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.800137 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.801671 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.802573 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.805436 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.806825 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-dpwlm"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.808114 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.809167 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fg658"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.812132 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.820825 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.823397 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k2b4r"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.824960 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.825668 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.826778 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.828110 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jdvbk"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.829546 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-smk4x"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.831134 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-92rk6"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.832027 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.833271 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.834363 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bjl65"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.835364 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl"] Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.843381 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.878349 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59pwc\" (UniqueName: \"kubernetes.io/projected/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-kube-api-access-59pwc\") pod \"controller-manager-879f6c89f-6wlqk\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.896912 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crxbp\" (UniqueName: \"kubernetes.io/projected/f1783a29-c599-40fe-a075-32732bb082db-kube-api-access-crxbp\") pod \"apiserver-7bbb656c7d-8k9cq\" (UID: \"f1783a29-c599-40fe-a075-32732bb082db\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.918896 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljd6p\" (UniqueName: \"kubernetes.io/projected/5279cafe-fc2e-4037-8841-ac430e0903b6-kube-api-access-ljd6p\") pod \"apiserver-76f77b778f-fg658\" (UID: \"5279cafe-fc2e-4037-8841-ac430e0903b6\") " pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.940643 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22grw\" (UniqueName: \"kubernetes.io/projected/f2de8f09-187b-4b8f-9799-321d14745677-kube-api-access-22grw\") pod \"openshift-apiserver-operator-796bbdcf4f-pfjsx\" (UID: \"f2de8f09-187b-4b8f-9799-321d14745677\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.942985 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.963818 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 08:40:18 crc kubenswrapper[4573]: I1203 08:40:18.984472 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.017474 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz9pv\" (UniqueName: \"kubernetes.io/projected/3e8a1377-41c2-4b26-8bd1-918fdde3d19a-kube-api-access-hz9pv\") pod \"console-operator-58897d9998-vphhd\" (UID: \"3e8a1377-41c2-4b26-8bd1-918fdde3d19a\") " pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.023743 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.036759 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.050421 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.050755 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.064381 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.064427 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.074629 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.083598 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.104834 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" event={"ID":"7807daee-9b44-4ebd-aab9-afaf91fb9746","Type":"ContainerStarted","Data":"a73f148ed48362e0eb5526e9fca4b85ae5843b78878701f0225b9f62260052c9"} Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.106541 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.126576 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.137077 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.169341 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67d2h\" (UniqueName: \"kubernetes.io/projected/cc4dd1dc-155a-4278-8382-cbe5364f80eb-kube-api-access-67d2h\") pod \"etcd-operator-b45778765-2w8js\" (UID: \"cc4dd1dc-155a-4278-8382-cbe5364f80eb\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.175255 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.184275 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bpl4\" (UniqueName: \"kubernetes.io/projected/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-kube-api-access-2bpl4\") pod \"console-f9d7485db-dpwlm\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.200152 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqgrt\" (UniqueName: \"kubernetes.io/projected/9b31f7fa-c98a-45e3-934e-94e70d254fca-kube-api-access-vqgrt\") pod \"cluster-samples-operator-665b6dd947-sjs5t\" (UID: \"9b31f7fa-c98a-45e3-934e-94e70d254fca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.205956 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.217783 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.225569 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.244854 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.271190 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.283757 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.309196 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6wlqk"] Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.328585 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.335890 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.343547 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.343770 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.363509 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.384228 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.403685 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.404608 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.424856 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vphhd"] Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.425993 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.444647 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.466391 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.494985 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.506433 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.544405 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.546740 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn2q5\" (UniqueName: \"kubernetes.io/projected/8b5ee846-009f-4feb-a61b-a413aee18c4d-kube-api-access-xn2q5\") pod \"openshift-config-operator-7777fb866f-l7qrz\" (UID: \"8b5ee846-009f-4feb-a61b-a413aee18c4d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.565711 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.580102 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq"] Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.583216 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx"] Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.589712 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.607033 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.623968 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.629220 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fg658"] Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.643029 4573 request.go:700] Waited for 1.00180113s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/configmaps?fieldSelector=metadata.name%3Dkube-apiserver-operator-config&limit=500&resourceVersion=0 Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.646456 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.658929 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-dpwlm"] Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.680805 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh52v\" (UniqueName: \"kubernetes.io/projected/284d9db2-04e5-4c39-b4cf-6f001de356c4-kube-api-access-bh52v\") pod \"machine-api-operator-5694c8668f-8qfqs\" (UID: \"284d9db2-04e5-4c39-b4cf-6f001de356c4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.683394 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.686983 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.697840 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t"] Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.704742 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.716915 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.724181 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.744475 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.763430 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.783674 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.804721 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.837138 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8p4l\" (UniqueName: \"kubernetes.io/projected/ee739812-3e92-42ff-a39d-fcea79a3164c-kube-api-access-n8p4l\") pod \"route-controller-manager-6576b87f9c-bzmrn\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.857425 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfc8w\" (UniqueName: \"kubernetes.io/projected/fef5501f-3c10-40c0-8f37-0383d15eb68f-kube-api-access-hfc8w\") pod \"dns-operator-744455d44c-vzjzb\" (UID: \"fef5501f-3c10-40c0-8f37-0383d15eb68f\") " pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.863915 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.883965 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.902832 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.902830 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.923398 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.943825 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 08:40:19 crc kubenswrapper[4573]: W1203 08:40:19.966350 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e8a1377_41c2_4b26_8bd1_918fdde3d19a.slice/crio-ae4102b16d681599ad0f541abcc5979f7405f34a9b90b5c698c9ee19f63b9b43 WatchSource:0}: Error finding container ae4102b16d681599ad0f541abcc5979f7405f34a9b90b5c698c9ee19f63b9b43: Status 404 returned error can't find the container with id ae4102b16d681599ad0f541abcc5979f7405f34a9b90b5c698c9ee19f63b9b43 Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.967531 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.977936 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" Dec 03 08:40:19 crc kubenswrapper[4573]: W1203 08:40:19.982861 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5279cafe_fc2e_4037_8841_ac430e0903b6.slice/crio-fa2868d5cf98970a55f72463fe02cf15a4829673d45930979b3aba328e29fa14 WatchSource:0}: Error finding container fa2868d5cf98970a55f72463fe02cf15a4829673d45930979b3aba328e29fa14: Status 404 returned error can't find the container with id fa2868d5cf98970a55f72463fe02cf15a4829673d45930979b3aba328e29fa14 Dec 03 08:40:19 crc kubenswrapper[4573]: I1203 08:40:19.983665 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.020720 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac517135-70a0-4f3d-b19c-9db51f67b2fc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.020805 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.020830 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.020864 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.020899 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vmwt\" (UniqueName: \"kubernetes.io/projected/b3c8277f-eef7-4121-8b4e-ed0df8deda4b-kube-api-access-6vmwt\") pod \"downloads-7954f5f757-b92rg\" (UID: \"b3c8277f-eef7-4121-8b4e-ed0df8deda4b\") " pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.020944 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6css7\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-kube-api-access-6css7\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.020967 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.021012 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.021037 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-tls\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.021118 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.021156 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/468b20da-e9e1-43e5-a265-54d131b1b66a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.021190 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.021215 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.021236 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-dir\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.021258 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znhgv\" (UniqueName: \"kubernetes.io/projected/76718a7d-3779-4ff8-a60d-5bef0372d0a2-kube-api-access-znhgv\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.022010 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:20.521987128 +0000 UTC m=+141.090366447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.022461 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-policies\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.022561 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffp97\" (UniqueName: \"kubernetes.io/projected/ac517135-70a0-4f3d-b19c-9db51f67b2fc-kube-api-access-ffp97\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.022745 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac517135-70a0-4f3d-b19c-9db51f67b2fc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.023015 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76718a7d-3779-4ff8-a60d-5bef0372d0a2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.023264 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf78f\" (UniqueName: \"kubernetes.io/projected/2cf8e819-4c7b-41c6-984c-871a6ac87428-kube-api-access-gf78f\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.023388 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76718a7d-3779-4ff8-a60d-5bef0372d0a2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.023493 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.023869 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.023954 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-certificates\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.024000 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-trusted-ca\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.024015 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.024040 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-bound-sa-token\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.024074 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac517135-70a0-4f3d-b19c-9db51f67b2fc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.024290 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.024599 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/468b20da-e9e1-43e5-a265-54d131b1b66a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.025610 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.043667 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.070790 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.091493 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.103712 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125365 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125634 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9ba983f3-7837-46f0-b5ff-b8277252e43c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125659 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxz6w\" (UniqueName: \"kubernetes.io/projected/54b30813-ea73-4e52-b6ac-c2d1eda3da53-kube-api-access-kxz6w\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125695 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7175f33-8e75-469b-a085-20a7553e2b27-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125713 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/468b20da-e9e1-43e5-a265-54d131b1b66a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125731 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-bound-sa-token\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125747 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac517135-70a0-4f3d-b19c-9db51f67b2fc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125772 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-default-certificate\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125793 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4932282a-3036-4392-ace4-1707a321268a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125815 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nftnc\" (UniqueName: \"kubernetes.io/projected/6c9346bd-44d0-47ea-9d16-594259b8a867-kube-api-access-nftnc\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125847 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c10cae9-fe93-4134-ae2f-7e242201dc5f-secret-volume\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125869 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vmwt\" (UniqueName: \"kubernetes.io/projected/b3c8277f-eef7-4121-8b4e-ed0df8deda4b-kube-api-access-6vmwt\") pod \"downloads-7954f5f757-b92rg\" (UID: \"b3c8277f-eef7-4121-8b4e-ed0df8deda4b\") " pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125894 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f7280fb-cb08-4761-9d89-2f356d797c89-serving-cert\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125915 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4jxgf\" (UID: \"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125938 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125959 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.125978 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwwgk\" (UniqueName: \"kubernetes.io/projected/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-kube-api-access-vwwgk\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126009 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-tls\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126027 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c9346bd-44d0-47ea-9d16-594259b8a867-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126042 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmxtr\" (UniqueName: \"kubernetes.io/projected/db0d1452-dca7-4745-9f61-6ae4ae162a0f-kube-api-access-qmxtr\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126077 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85a90ac-b6d5-41c3-808f-eb66020ab534-config\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126094 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e85a90ac-b6d5-41c3-808f-eb66020ab534-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126110 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-service-ca-bundle\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126126 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-socket-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126141 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/468b20da-e9e1-43e5-a265-54d131b1b66a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126156 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/715be642-2503-4da6-9caf-05ff95d26fc3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pxklw\" (UID: \"715be642-2503-4da6-9caf-05ff95d26fc3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126174 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126189 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126205 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbwfh\" (UniqueName: \"kubernetes.io/projected/fd99d655-f2f2-451b-a058-2a70e262fb2d-kube-api-access-mbwfh\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126226 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znhgv\" (UniqueName: \"kubernetes.io/projected/76718a7d-3779-4ff8-a60d-5bef0372d0a2-kube-api-access-znhgv\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126243 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-plugins-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126258 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-registration-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126275 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nts4p\" (UniqueName: \"kubernetes.io/projected/74479bf9-6a28-41a1-9a70-dca91bfbb855-kube-api-access-nts4p\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126291 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4932282a-3036-4392-ace4-1707a321268a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126318 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac517135-70a0-4f3d-b19c-9db51f67b2fc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126340 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7175f33-8e75-469b-a085-20a7553e2b27-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126357 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126371 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg9ff\" (UniqueName: \"kubernetes.io/projected/ca368157-8579-4618-8031-a793b862bacd-kube-api-access-cg9ff\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126392 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/74479bf9-6a28-41a1-9a70-dca91bfbb855-metrics-tls\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126410 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd99d655-f2f2-451b-a058-2a70e262fb2d-config\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126439 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76718a7d-3779-4ff8-a60d-5bef0372d0a2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126462 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htg6b\" (UniqueName: \"kubernetes.io/projected/2c10cae9-fe93-4134-ae2f-7e242201dc5f-kube-api-access-htg6b\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126484 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126505 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4932282a-3036-4392-ace4-1707a321268a-config\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126534 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-certificates\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126551 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/54b30813-ea73-4e52-b6ac-c2d1eda3da53-node-bootstrap-token\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126566 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ca368157-8579-4618-8031-a793b862bacd-metrics-tls\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126584 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-trusted-ca\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126599 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4m79\" (UniqueName: \"kubernetes.io/projected/192416c5-8025-4dd5-840b-9954ce7d009b-kube-api-access-v4m79\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mk94\" (UID: \"192416c5-8025-4dd5-840b-9954ce7d009b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126614 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxmzp\" (UniqueName: \"kubernetes.io/projected/9ba983f3-7837-46f0-b5ff-b8277252e43c-kube-api-access-wxmzp\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126630 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4dn5\" (UniqueName: \"kubernetes.io/projected/76462ae4-1687-4489-b12d-0e6817753e41-kube-api-access-h4dn5\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126648 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126667 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126683 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac517135-70a0-4f3d-b19c-9db51f67b2fc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126701 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01a2ef90-923e-4843-91bd-5e367c232a0c-apiservice-cert\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126715 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/73472804-4ec7-4d1a-9a34-c6cb9e11b709-signing-key\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126732 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5dlv\" (UniqueName: \"kubernetes.io/projected/715be642-2503-4da6-9caf-05ff95d26fc3-kube-api-access-p5dlv\") pod \"multus-admission-controller-857f4d67dd-pxklw\" (UID: \"715be642-2503-4da6-9caf-05ff95d26fc3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126766 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/192416c5-8025-4dd5-840b-9954ce7d009b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mk94\" (UID: \"192416c5-8025-4dd5-840b-9954ce7d009b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126793 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/54b30813-ea73-4e52-b6ac-c2d1eda3da53-certs\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126836 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126853 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126879 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126895 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76462ae4-1687-4489-b12d-0e6817753e41-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126909 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf-cert\") pod \"ingress-canary-bjl65\" (UID: \"d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf\") " pod="openshift-ingress-canary/ingress-canary-bjl65" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126924 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42f8h\" (UniqueName: \"kubernetes.io/projected/f6e10744-6f93-416a-a6aa-4a7a07c76eff-kube-api-access-42f8h\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126941 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6css7\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-kube-api-access-6css7\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126965 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-stats-auth\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126982 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.126999 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c10cae9-fe93-4134-ae2f-7e242201dc5f-config-volume\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127022 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c9346bd-44d0-47ea-9d16-594259b8a867-proxy-tls\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127038 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fl8d\" (UniqueName: \"kubernetes.io/projected/a44990ef-b3ba-4db7-98b2-7c6100a3acae-kube-api-access-5fl8d\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127057 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a7efaa-3253-40d5-9e1e-59482378cd8c-service-ca-bundle\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127086 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127101 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/73472804-4ec7-4d1a-9a34-c6cb9e11b709-signing-cabundle\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127116 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dzn5\" (UniqueName: \"kubernetes.io/projected/aef07f9d-ddd0-49a3-b1cb-7841f061a24e-kube-api-access-9dzn5\") pod \"migrator-59844c95c7-7lnsg\" (UID: \"aef07f9d-ddd0-49a3-b1cb-7841f061a24e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127130 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-dir\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127155 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfp4l\" (UniqueName: \"kubernetes.io/projected/01a2ef90-923e-4843-91bd-5e367c232a0c-kube-api-access-jfp4l\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127172 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6e10744-6f93-416a-a6aa-4a7a07c76eff-srv-cert\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127187 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74479bf9-6a28-41a1-9a70-dca91bfbb855-trusted-ca\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127203 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-policies\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127218 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6c9346bd-44d0-47ea-9d16-594259b8a867-images\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127260 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffp97\" (UniqueName: \"kubernetes.io/projected/ac517135-70a0-4f3d-b19c-9db51f67b2fc-kube-api-access-ffp97\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127277 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01a2ef90-923e-4843-91bd-5e367c232a0c-webhook-cert\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127293 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jdr2\" (UniqueName: \"kubernetes.io/projected/5f7280fb-cb08-4761-9d89-2f356d797c89-kube-api-access-8jdr2\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127309 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76462ae4-1687-4489-b12d-0e6817753e41-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127326 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76718a7d-3779-4ff8-a60d-5bef0372d0a2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127341 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01a2ef90-923e-4843-91bd-5e367c232a0c-tmpfs\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127380 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-config\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127397 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf78f\" (UniqueName: \"kubernetes.io/projected/2cf8e819-4c7b-41c6-984c-871a6ac87428-kube-api-access-gf78f\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127412 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kw49\" (UniqueName: \"kubernetes.io/projected/73472804-4ec7-4d1a-9a34-c6cb9e11b709-kube-api-access-8kw49\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127427 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-metrics-certs\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127443 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg5nm\" (UniqueName: \"kubernetes.io/projected/0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb-kube-api-access-kg5nm\") pod \"package-server-manager-789f6589d5-4jxgf\" (UID: \"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127460 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-mountpoint-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127477 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz5zp\" (UniqueName: \"kubernetes.io/projected/d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf-kube-api-access-cz5zp\") pod \"ingress-canary-bjl65\" (UID: \"d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf\") " pod="openshift-ingress-canary/ingress-canary-bjl65" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127492 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-csi-data-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127525 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e85a90ac-b6d5-41c3-808f-eb66020ab534-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127544 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127560 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j59p\" (UniqueName: \"kubernetes.io/projected/f5a7efaa-3253-40d5-9e1e-59482378cd8c-kube-api-access-4j59p\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127575 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd99d655-f2f2-451b-a058-2a70e262fb2d-serving-cert\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127591 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-srv-cert\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127609 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127625 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca368157-8579-4618-8031-a793b862bacd-config-volume\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127641 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7175f33-8e75-469b-a085-20a7553e2b27-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127656 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6e10744-6f93-416a-a6aa-4a7a07c76eff-profile-collector-cert\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127671 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ba983f3-7837-46f0-b5ff-b8277252e43c-proxy-tls\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.127688 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74479bf9-6a28-41a1-9a70-dca91bfbb855-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.127800 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:20.627782878 +0000 UTC m=+141.196162137 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.128653 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/468b20da-e9e1-43e5-a265-54d131b1b66a-ca-trust-extracted\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.129291 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.131032 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac517135-70a0-4f3d-b19c-9db51f67b2fc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.132909 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.141593 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-certificates\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.142353 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.142424 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-dir\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.143787 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.144329 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.145132 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-trusted-ca\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.146991 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac517135-70a0-4f3d-b19c-9db51f67b2fc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.147544 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-policies\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.150094 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76718a7d-3779-4ff8-a60d-5bef0372d0a2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.150442 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.151147 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.152223 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.152496 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.152732 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.152981 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76718a7d-3779-4ff8-a60d-5bef0372d0a2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.153429 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.154083 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.154302 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.155580 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/468b20da-e9e1-43e5-a265-54d131b1b66a-installation-pull-secrets\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.163971 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-tls\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.163993 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.187658 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.189146 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dpwlm" event={"ID":"bdc5518a-70d7-4f7f-bb7a-3320cb61f607","Type":"ContainerStarted","Data":"a450a21c80f33dcd780dd699f6b999e90b55e9c97d97b5019e1a10cd0b153ab0"} Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.191088 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" event={"ID":"f1783a29-c599-40fe-a075-32732bb082db","Type":"ContainerStarted","Data":"49a33f4556ba24e3509d6df5ddf9a7ca70f4d69ad9cd9f2300258996934d48d6"} Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.200208 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" event={"ID":"7807daee-9b44-4ebd-aab9-afaf91fb9746","Type":"ContainerStarted","Data":"2b19608d27108b9e39bc5778dacd98944edd272c1935d0bacb5410ea18e03bb9"} Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.201516 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vphhd" event={"ID":"3e8a1377-41c2-4b26-8bd1-918fdde3d19a","Type":"ContainerStarted","Data":"ae4102b16d681599ad0f541abcc5979f7405f34a9b90b5c698c9ee19f63b9b43"} Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.204474 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.224012 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228341 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htg6b\" (UniqueName: \"kubernetes.io/projected/2c10cae9-fe93-4134-ae2f-7e242201dc5f-kube-api-access-htg6b\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228387 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228406 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4932282a-3036-4392-ace4-1707a321268a-config\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228422 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/54b30813-ea73-4e52-b6ac-c2d1eda3da53-node-bootstrap-token\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228455 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ca368157-8579-4618-8031-a793b862bacd-metrics-tls\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228472 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxmzp\" (UniqueName: \"kubernetes.io/projected/9ba983f3-7837-46f0-b5ff-b8277252e43c-kube-api-access-wxmzp\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228490 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4dn5\" (UniqueName: \"kubernetes.io/projected/76462ae4-1687-4489-b12d-0e6817753e41-kube-api-access-h4dn5\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228506 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4m79\" (UniqueName: \"kubernetes.io/projected/192416c5-8025-4dd5-840b-9954ce7d009b-kube-api-access-v4m79\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mk94\" (UID: \"192416c5-8025-4dd5-840b-9954ce7d009b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228541 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01a2ef90-923e-4843-91bd-5e367c232a0c-apiservice-cert\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228558 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/73472804-4ec7-4d1a-9a34-c6cb9e11b709-signing-key\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228574 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5dlv\" (UniqueName: \"kubernetes.io/projected/715be642-2503-4da6-9caf-05ff95d26fc3-kube-api-access-p5dlv\") pod \"multus-admission-controller-857f4d67dd-pxklw\" (UID: \"715be642-2503-4da6-9caf-05ff95d26fc3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228610 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/54b30813-ea73-4e52-b6ac-c2d1eda3da53-certs\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228631 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/192416c5-8025-4dd5-840b-9954ce7d009b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mk94\" (UID: \"192416c5-8025-4dd5-840b-9954ce7d009b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228647 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76462ae4-1687-4489-b12d-0e6817753e41-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228663 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf-cert\") pod \"ingress-canary-bjl65\" (UID: \"d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf\") " pod="openshift-ingress-canary/ingress-canary-bjl65" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228698 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42f8h\" (UniqueName: \"kubernetes.io/projected/f6e10744-6f93-416a-a6aa-4a7a07c76eff-kube-api-access-42f8h\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228720 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-stats-auth\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228736 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c10cae9-fe93-4134-ae2f-7e242201dc5f-config-volume\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228773 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c9346bd-44d0-47ea-9d16-594259b8a867-proxy-tls\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228790 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fl8d\" (UniqueName: \"kubernetes.io/projected/a44990ef-b3ba-4db7-98b2-7c6100a3acae-kube-api-access-5fl8d\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228806 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a7efaa-3253-40d5-9e1e-59482378cd8c-service-ca-bundle\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228824 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228870 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/73472804-4ec7-4d1a-9a34-c6cb9e11b709-signing-cabundle\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228888 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dzn5\" (UniqueName: \"kubernetes.io/projected/aef07f9d-ddd0-49a3-b1cb-7841f061a24e-kube-api-access-9dzn5\") pod \"migrator-59844c95c7-7lnsg\" (UID: \"aef07f9d-ddd0-49a3-b1cb-7841f061a24e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228923 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfp4l\" (UniqueName: \"kubernetes.io/projected/01a2ef90-923e-4843-91bd-5e367c232a0c-kube-api-access-jfp4l\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228946 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6e10744-6f93-416a-a6aa-4a7a07c76eff-srv-cert\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228961 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74479bf9-6a28-41a1-9a70-dca91bfbb855-trusted-ca\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.228977 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6c9346bd-44d0-47ea-9d16-594259b8a867-images\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229013 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01a2ef90-923e-4843-91bd-5e367c232a0c-webhook-cert\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229030 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jdr2\" (UniqueName: \"kubernetes.io/projected/5f7280fb-cb08-4761-9d89-2f356d797c89-kube-api-access-8jdr2\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229065 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76462ae4-1687-4489-b12d-0e6817753e41-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229081 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01a2ef90-923e-4843-91bd-5e367c232a0c-tmpfs\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229096 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-config\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229112 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-metrics-certs\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229146 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg5nm\" (UniqueName: \"kubernetes.io/projected/0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb-kube-api-access-kg5nm\") pod \"package-server-manager-789f6589d5-4jxgf\" (UID: \"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229163 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-mountpoint-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229179 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz5zp\" (UniqueName: \"kubernetes.io/projected/d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf-kube-api-access-cz5zp\") pod \"ingress-canary-bjl65\" (UID: \"d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf\") " pod="openshift-ingress-canary/ingress-canary-bjl65" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229199 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kw49\" (UniqueName: \"kubernetes.io/projected/73472804-4ec7-4d1a-9a34-c6cb9e11b709-kube-api-access-8kw49\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229231 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-csi-data-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229246 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e85a90ac-b6d5-41c3-808f-eb66020ab534-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229262 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j59p\" (UniqueName: \"kubernetes.io/projected/f5a7efaa-3253-40d5-9e1e-59482378cd8c-kube-api-access-4j59p\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229277 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd99d655-f2f2-451b-a058-2a70e262fb2d-serving-cert\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229309 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-srv-cert\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229325 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca368157-8579-4618-8031-a793b862bacd-config-volume\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229342 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7175f33-8e75-469b-a085-20a7553e2b27-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229374 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6e10744-6f93-416a-a6aa-4a7a07c76eff-profile-collector-cert\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229414 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ba983f3-7837-46f0-b5ff-b8277252e43c-proxy-tls\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229430 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74479bf9-6a28-41a1-9a70-dca91bfbb855-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229465 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9ba983f3-7837-46f0-b5ff-b8277252e43c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229481 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxz6w\" (UniqueName: \"kubernetes.io/projected/54b30813-ea73-4e52-b6ac-c2d1eda3da53-kube-api-access-kxz6w\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229504 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7175f33-8e75-469b-a085-20a7553e2b27-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229542 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-default-certificate\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229558 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4932282a-3036-4392-ace4-1707a321268a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229574 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nftnc\" (UniqueName: \"kubernetes.io/projected/6c9346bd-44d0-47ea-9d16-594259b8a867-kube-api-access-nftnc\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229615 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c10cae9-fe93-4134-ae2f-7e242201dc5f-secret-volume\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229631 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f7280fb-cb08-4761-9d89-2f356d797c89-serving-cert\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229653 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4jxgf\" (UID: \"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229668 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwwgk\" (UniqueName: \"kubernetes.io/projected/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-kube-api-access-vwwgk\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229700 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229715 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c9346bd-44d0-47ea-9d16-594259b8a867-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229731 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmxtr\" (UniqueName: \"kubernetes.io/projected/db0d1452-dca7-4745-9f61-6ae4ae162a0f-kube-api-access-qmxtr\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229766 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229784 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85a90ac-b6d5-41c3-808f-eb66020ab534-config\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229799 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e85a90ac-b6d5-41c3-808f-eb66020ab534-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229815 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-service-ca-bundle\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229846 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-socket-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229863 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/715be642-2503-4da6-9caf-05ff95d26fc3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pxklw\" (UID: \"715be642-2503-4da6-9caf-05ff95d26fc3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229881 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbwfh\" (UniqueName: \"kubernetes.io/projected/fd99d655-f2f2-451b-a058-2a70e262fb2d-kube-api-access-mbwfh\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229901 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-plugins-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229935 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-registration-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229952 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nts4p\" (UniqueName: \"kubernetes.io/projected/74479bf9-6a28-41a1-9a70-dca91bfbb855-kube-api-access-nts4p\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.229974 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4932282a-3036-4392-ace4-1707a321268a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.230014 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7175f33-8e75-469b-a085-20a7553e2b27-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.230030 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.230065 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg9ff\" (UniqueName: \"kubernetes.io/projected/ca368157-8579-4618-8031-a793b862bacd-kube-api-access-cg9ff\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.230081 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/74479bf9-6a28-41a1-9a70-dca91bfbb855-metrics-tls\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.230096 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd99d655-f2f2-451b-a058-2a70e262fb2d-config\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.230994 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd99d655-f2f2-451b-a058-2a70e262fb2d-config\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.231162 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-csi-data-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.231802 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4932282a-3036-4392-ace4-1707a321268a-config\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.233550 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:20.733531697 +0000 UTC m=+141.301911026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.236207 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9ba983f3-7837-46f0-b5ff-b8277252e43c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.236590 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-plugins-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.236749 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c9346bd-44d0-47ea-9d16-594259b8a867-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.237466 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.237522 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85a90ac-b6d5-41c3-808f-eb66020ab534-config\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.240695 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd99d655-f2f2-451b-a058-2a70e262fb2d-serving-cert\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.241430 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7175f33-8e75-469b-a085-20a7553e2b27-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.243893 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7175f33-8e75-469b-a085-20a7553e2b27-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.244001 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-registration-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.245455 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6e10744-6f93-416a-a6aa-4a7a07c76eff-profile-collector-cert\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.246861 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-service-ca-bundle\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.246937 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-socket-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.247331 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e85a90ac-b6d5-41c3-808f-eb66020ab534-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.247838 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.248541 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/74479bf9-6a28-41a1-9a70-dca91bfbb855-metrics-tls\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.249552 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ba983f3-7837-46f0-b5ff-b8277252e43c-proxy-tls\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.249839 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/01a2ef90-923e-4843-91bd-5e367c232a0c-tmpfs\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.250946 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74479bf9-6a28-41a1-9a70-dca91bfbb855-trusted-ca\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.251387 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6c9346bd-44d0-47ea-9d16-594259b8a867-images\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.251926 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/76462ae4-1687-4489-b12d-0e6817753e41-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.252076 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.252586 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f7280fb-cb08-4761-9d89-2f356d797c89-config\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.254845 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6e10744-6f93-416a-a6aa-4a7a07c76eff-srv-cert\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.255942 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a7efaa-3253-40d5-9e1e-59482378cd8c-service-ca-bundle\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.258078 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c9346bd-44d0-47ea-9d16-594259b8a867-proxy-tls\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.258435 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/715be642-2503-4da6-9caf-05ff95d26fc3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pxklw\" (UID: \"715be642-2503-4da6-9caf-05ff95d26fc3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.258623 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/db0d1452-dca7-4745-9f61-6ae4ae162a0f-mountpoint-dir\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.259979 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-metrics-certs\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.260343 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4932282a-3036-4392-ace4-1707a321268a-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.261072 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f7280fb-cb08-4761-9d89-2f356d797c89-serving-cert\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.261662 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.262205 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-default-certificate\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.262622 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.262759 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c10cae9-fe93-4134-ae2f-7e242201dc5f-secret-volume\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.263261 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f5a7efaa-3253-40d5-9e1e-59482378cd8c-stats-auth\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.263630 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.265248 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/76462ae4-1687-4489-b12d-0e6817753e41-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.270592 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/192416c5-8025-4dd5-840b-9954ce7d009b-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mk94\" (UID: \"192416c5-8025-4dd5-840b-9954ce7d009b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.283961 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" event={"ID":"a77aa3c0-2caf-48c1-a771-d374b1d3e5de","Type":"ContainerStarted","Data":"8b8415d7dbdb09ecbab0ec26b0cba456aee7e293bbe4b381d525a1d79ad19a60"} Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.285238 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.287925 4573 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6wlqk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.287965 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" podUID="a77aa3c0-2caf-48c1-a771-d374b1d3e5de" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.291874 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" event={"ID":"9b31f7fa-c98a-45e3-934e-94e70d254fca","Type":"ContainerStarted","Data":"7310e2ce219299a7862502c2f1f39bbfeed4bf5732b8162352ce48d043cd06ca"} Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.294310 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.364384 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.364962 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:20.864948238 +0000 UTC m=+141.433327497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.365649 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fg658" event={"ID":"5279cafe-fc2e-4037-8841-ac430e0903b6","Type":"ContainerStarted","Data":"fa2868d5cf98970a55f72463fe02cf15a4829673d45930979b3aba328e29fa14"} Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.368149 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" event={"ID":"f2de8f09-187b-4b8f-9799-321d14745677","Type":"ContainerStarted","Data":"a66fcd09caa9cd8b2af2af98cc13258101b7b1399b8a3a8e971bb76e6ff15aa6"} Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.370312 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-srv-cert\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.370903 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.371152 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.371215 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.376161 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01a2ef90-923e-4843-91bd-5e367c232a0c-apiservice-cert\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.377819 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01a2ef90-923e-4843-91bd-5e367c232a0c-webhook-cert\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.378652 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4jxgf\" (UID: \"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.382963 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.432008 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.436373 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.439956 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/73472804-4ec7-4d1a-9a34-c6cb9e11b709-signing-cabundle\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.444240 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/73472804-4ec7-4d1a-9a34-c6cb9e11b709-signing-key\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.445761 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.465932 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.466876 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:20.966863746 +0000 UTC m=+141.535243005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.468470 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.470752 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c10cae9-fe93-4134-ae2f-7e242201dc5f-config-volume\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.479525 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vzjzb"] Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.486184 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.515036 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.531397 4573 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.538801 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz"] Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.545387 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.564434 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.567285 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.567804 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.067787609 +0000 UTC m=+141.636166868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.586811 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.598107 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf-cert\") pod \"ingress-canary-bjl65\" (UID: \"d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf\") " pod="openshift-ingress-canary/ingress-canary-bjl65" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.609260 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.628582 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.643618 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.661563 4573 request.go:700] Waited for 1.875194047s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.663943 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.668269 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2w8js"] Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.668743 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.669268 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.169255335 +0000 UTC m=+141.737634604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.678584 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/54b30813-ea73-4e52-b6ac-c2d1eda3da53-node-bootstrap-token\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.685262 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.698488 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/54b30813-ea73-4e52-b6ac-c2d1eda3da53-certs\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.700136 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn"] Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.703599 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.712914 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ca368157-8579-4618-8031-a793b862bacd-config-volume\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.724429 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.727577 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8qfqs"] Dec 03 08:40:20 crc kubenswrapper[4573]: W1203 08:40:20.739816 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod284d9db2_04e5_4c39_b4cf_6f001de356c4.slice/crio-689c1f4f36085111b43822ed50bf78502a5fddd3521f18ddf883f7e4531536ff WatchSource:0}: Error finding container 689c1f4f36085111b43822ed50bf78502a5fddd3521f18ddf883f7e4531536ff: Status 404 returned error can't find the container with id 689c1f4f36085111b43822ed50bf78502a5fddd3521f18ddf883f7e4531536ff Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.742562 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.747982 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ca368157-8579-4618-8031-a793b862bacd-metrics-tls\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.769428 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.770184 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.270164827 +0000 UTC m=+141.838544086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.777511 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-bound-sa-token\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.818228 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffp97\" (UniqueName: \"kubernetes.io/projected/ac517135-70a0-4f3d-b19c-9db51f67b2fc-kube-api-access-ffp97\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.840117 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf78f\" (UniqueName: \"kubernetes.io/projected/2cf8e819-4c7b-41c6-984c-871a6ac87428-kube-api-access-gf78f\") pod \"oauth-openshift-558db77b4-m7kqr\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.856645 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znhgv\" (UniqueName: \"kubernetes.io/projected/76718a7d-3779-4ff8-a60d-5bef0372d0a2-kube-api-access-znhgv\") pod \"openshift-controller-manager-operator-756b6f6bc6-cj8dq\" (UID: \"76718a7d-3779-4ff8-a60d-5bef0372d0a2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.861140 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vmwt\" (UniqueName: \"kubernetes.io/projected/b3c8277f-eef7-4121-8b4e-ed0df8deda4b-kube-api-access-6vmwt\") pod \"downloads-7954f5f757-b92rg\" (UID: \"b3c8277f-eef7-4121-8b4e-ed0df8deda4b\") " pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.868348 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.871564 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.871886 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.37187291 +0000 UTC m=+141.940252169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.877819 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6css7\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-kube-api-access-6css7\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.896494 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac517135-70a0-4f3d-b19c-9db51f67b2fc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2rwtg\" (UID: \"ac517135-70a0-4f3d-b19c-9db51f67b2fc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.902022 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.918039 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htg6b\" (UniqueName: \"kubernetes.io/projected/2c10cae9-fe93-4134-ae2f-7e242201dc5f-kube-api-access-htg6b\") pod \"collect-profiles-29412510-r8ngs\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.948946 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxmzp\" (UniqueName: \"kubernetes.io/projected/9ba983f3-7837-46f0-b5ff-b8277252e43c-kube-api-access-wxmzp\") pod \"machine-config-controller-84d6567774-pj2pz\" (UID: \"9ba983f3-7837-46f0-b5ff-b8277252e43c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.973262 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:20 crc kubenswrapper[4573]: E1203 08:40:20.975083 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.475064351 +0000 UTC m=+142.043443600 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.976631 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4dn5\" (UniqueName: \"kubernetes.io/projected/76462ae4-1687-4489-b12d-0e6817753e41-kube-api-access-h4dn5\") pod \"kube-storage-version-migrator-operator-b67b599dd-d8w5j\" (UID: \"76462ae4-1687-4489-b12d-0e6817753e41\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.981804 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" Dec 03 08:40:20 crc kubenswrapper[4573]: I1203 08:40:20.988014 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4m79\" (UniqueName: \"kubernetes.io/projected/192416c5-8025-4dd5-840b-9954ce7d009b-kube-api-access-v4m79\") pod \"control-plane-machine-set-operator-78cbb6b69f-7mk94\" (UID: \"192416c5-8025-4dd5-840b-9954ce7d009b\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.009628 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5dlv\" (UniqueName: \"kubernetes.io/projected/715be642-2503-4da6-9caf-05ff95d26fc3-kube-api-access-p5dlv\") pod \"multus-admission-controller-857f4d67dd-pxklw\" (UID: \"715be642-2503-4da6-9caf-05ff95d26fc3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.026334 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwwgk\" (UniqueName: \"kubernetes.io/projected/43fd56dc-b8c0-4d05-8ca2-9314ce4e6059-kube-api-access-vwwgk\") pod \"olm-operator-6b444d44fb-z2gvw\" (UID: \"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.030107 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.062528 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.072034 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7175f33-8e75-469b-a085-20a7553e2b27-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zwvkr\" (UID: \"c7175f33-8e75-469b-a085-20a7553e2b27\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.076118 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.076485 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.576469317 +0000 UTC m=+142.144848586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.078785 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxz6w\" (UniqueName: \"kubernetes.io/projected/54b30813-ea73-4e52-b6ac-c2d1eda3da53-kube-api-access-kxz6w\") pod \"machine-config-server-4twfh\" (UID: \"54b30813-ea73-4e52-b6ac-c2d1eda3da53\") " pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.108482 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j59p\" (UniqueName: \"kubernetes.io/projected/f5a7efaa-3253-40d5-9e1e-59482378cd8c-kube-api-access-4j59p\") pod \"router-default-5444994796-8tpmr\" (UID: \"f5a7efaa-3253-40d5-9e1e-59482378cd8c\") " pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.108876 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4twfh" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.125338 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmxtr\" (UniqueName: \"kubernetes.io/projected/db0d1452-dca7-4745-9f61-6ae4ae162a0f-kube-api-access-qmxtr\") pod \"csi-hostpathplugin-smk4x\" (UID: \"db0d1452-dca7-4745-9f61-6ae4ae162a0f\") " pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.151108 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e85a90ac-b6d5-41c3-808f-eb66020ab534-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zt5bd\" (UID: \"e85a90ac-b6d5-41c3-808f-eb66020ab534\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.158622 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.169726 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nts4p\" (UniqueName: \"kubernetes.io/projected/74479bf9-6a28-41a1-9a70-dca91bfbb855-kube-api-access-nts4p\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.177361 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.177955 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.677924043 +0000 UTC m=+142.246303302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.191411 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.197991 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg9ff\" (UniqueName: \"kubernetes.io/projected/ca368157-8579-4618-8031-a793b862bacd-kube-api-access-cg9ff\") pod \"dns-default-92rk6\" (UID: \"ca368157-8579-4618-8031-a793b862bacd\") " pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.211402 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.226512 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.276133 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.277127 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nftnc\" (UniqueName: \"kubernetes.io/projected/6c9346bd-44d0-47ea-9d16-594259b8a867-kube-api-access-nftnc\") pod \"machine-config-operator-74547568cd-mr8km\" (UID: \"6c9346bd-44d0-47ea-9d16-594259b8a867\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.277444 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4932282a-3036-4392-ace4-1707a321268a-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-4j98x\" (UID: \"4932282a-3036-4392-ace4-1707a321268a\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.277773 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.278393 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.281190 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq"] Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.288866 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.295082 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.79504936 +0000 UTC m=+142.363428619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.304289 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.304814 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.312543 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fl8d\" (UniqueName: \"kubernetes.io/projected/a44990ef-b3ba-4db7-98b2-7c6100a3acae-kube-api-access-5fl8d\") pod \"marketplace-operator-79b997595-k2b4r\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.312990 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.316384 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jdr2\" (UniqueName: \"kubernetes.io/projected/5f7280fb-cb08-4761-9d89-2f356d797c89-kube-api-access-8jdr2\") pod \"authentication-operator-69f744f599-2ktws\" (UID: \"5f7280fb-cb08-4761-9d89-2f356d797c89\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.338105 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg5nm\" (UniqueName: \"kubernetes.io/projected/0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb-kube-api-access-kg5nm\") pod \"package-server-manager-789f6589d5-4jxgf\" (UID: \"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:21 crc kubenswrapper[4573]: W1203 08:40:21.342876 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76718a7d_3779_4ff8_a60d_5bef0372d0a2.slice/crio-d434c5961c35b8317b9bf7d8b292312172174a270a5d62766440ca9ebb5c058e WatchSource:0}: Error finding container d434c5961c35b8317b9bf7d8b292312172174a270a5d62766440ca9ebb5c058e: Status 404 returned error can't find the container with id d434c5961c35b8317b9bf7d8b292312172174a270a5d62766440ca9ebb5c058e Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.347102 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.351924 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/74479bf9-6a28-41a1-9a70-dca91bfbb855-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wd2xp\" (UID: \"74479bf9-6a28-41a1-9a70-dca91bfbb855\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.360411 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dzn5\" (UniqueName: \"kubernetes.io/projected/aef07f9d-ddd0-49a3-b1cb-7841f061a24e-kube-api-access-9dzn5\") pod \"migrator-59844c95c7-7lnsg\" (UID: \"aef07f9d-ddd0-49a3-b1cb-7841f061a24e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.368764 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfp4l\" (UniqueName: \"kubernetes.io/projected/01a2ef90-923e-4843-91bd-5e367c232a0c-kube-api-access-jfp4l\") pod \"packageserver-d55dfcdfc-4rx5b\" (UID: \"01a2ef90-923e-4843-91bd-5e367c232a0c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.372098 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz5zp\" (UniqueName: \"kubernetes.io/projected/d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf-kube-api-access-cz5zp\") pod \"ingress-canary-bjl65\" (UID: \"d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf\") " pod="openshift-ingress-canary/ingress-canary-bjl65" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.378602 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" event={"ID":"fef5501f-3c10-40c0-8f37-0383d15eb68f","Type":"ContainerStarted","Data":"36e816fab4d650315aa34f7770ccb33916e24dd95d2060f2d03f917cc13c9ef6"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.384976 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbwfh\" (UniqueName: \"kubernetes.io/projected/fd99d655-f2f2-451b-a058-2a70e262fb2d-kube-api-access-mbwfh\") pod \"service-ca-operator-777779d784-qdnvh\" (UID: \"fd99d655-f2f2-451b-a058-2a70e262fb2d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.395157 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-smk4x" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.400349 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bjl65" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.401509 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.402956 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:21.902490562 +0000 UTC m=+142.470869811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.403713 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" event={"ID":"284d9db2-04e5-4c39-b4cf-6f001de356c4","Type":"ContainerStarted","Data":"689c1f4f36085111b43822ed50bf78502a5fddd3521f18ddf883f7e4531536ff"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.412220 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4twfh" event={"ID":"54b30813-ea73-4e52-b6ac-c2d1eda3da53","Type":"ContainerStarted","Data":"496c4501dbfd7bbafbee861b680e64d038996d83cf237779d26e4e9fac2af003"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.415066 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42f8h\" (UniqueName: \"kubernetes.io/projected/f6e10744-6f93-416a-a6aa-4a7a07c76eff-kube-api-access-42f8h\") pod \"catalog-operator-68c6474976-z2tsl\" (UID: \"f6e10744-6f93-416a-a6aa-4a7a07c76eff\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.417650 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kw49\" (UniqueName: \"kubernetes.io/projected/73472804-4ec7-4d1a-9a34-c6cb9e11b709-kube-api-access-8kw49\") pod \"service-ca-9c57cc56f-jdvbk\" (UID: \"73472804-4ec7-4d1a-9a34-c6cb9e11b709\") " pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.418235 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.483307 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" event={"ID":"f2de8f09-187b-4b8f-9799-321d14745677","Type":"ContainerStarted","Data":"e0b567caa3a31c7a77635efa30a0210a2f248fef91fc004788ba0a4834645891"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.485460 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" event={"ID":"76718a7d-3779-4ff8-a60d-5bef0372d0a2","Type":"ContainerStarted","Data":"d434c5961c35b8317b9bf7d8b292312172174a270a5d62766440ca9ebb5c058e"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.503719 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.504006 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.00399516 +0000 UTC m=+142.572374419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.537787 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.538280 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.573386 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.574428 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.604757 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.605214 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.105194939 +0000 UTC m=+142.673574198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.611447 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.646680 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.647514 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.653085 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.764340 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.765360 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.265348042 +0000 UTC m=+142.833727301 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.775387 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m7kqr"] Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.776578 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dpwlm" event={"ID":"bdc5518a-70d7-4f7f-bb7a-3320cb61f607","Type":"ContainerStarted","Data":"c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.797383 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" event={"ID":"ee739812-3e92-42ff-a39d-fcea79a3164c","Type":"ContainerStarted","Data":"a7e972a7d2394f7a3753f501aeb3bd9f35263590fd328c1fee296d6c499de954"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.805527 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.806739 4573 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-bzmrn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.806770 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" podUID="ee739812-3e92-42ff-a39d-fcea79a3164c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.30:8443/healthz\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.816929 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" event={"ID":"8b5ee846-009f-4feb-a61b-a413aee18c4d","Type":"ContainerStarted","Data":"ebc877b220b1d7c06eeb885ae142ab895fabdd83712fcb2725ef8874742a3850"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.824709 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vphhd" event={"ID":"3e8a1377-41c2-4b26-8bd1-918fdde3d19a","Type":"ContainerStarted","Data":"1b6ed26cf47a1a93647afe19dfa0237e130e84a44f9c85cfb28bfb6b5de6fa03"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.825281 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.826487 4573 patch_prober.go:28] interesting pod/console-operator-58897d9998-vphhd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.826533 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-vphhd" podUID="3e8a1377-41c2-4b26-8bd1-918fdde3d19a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.860226 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" event={"ID":"a77aa3c0-2caf-48c1-a771-d374b1d3e5de","Type":"ContainerStarted","Data":"590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.860649 4573 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6wlqk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.860693 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" podUID="a77aa3c0-2caf-48c1-a771-d374b1d3e5de" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.865577 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.870759 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.370733612 +0000 UTC m=+142.939112871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.885823 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz"] Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.894746 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw"] Dec 03 08:40:21 crc kubenswrapper[4573]: W1203 08:40:21.907461 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5a7efaa_3253_40d5_9e1e_59482378cd8c.slice/crio-4192088e8b50de0b64d86c28705e817f1640d388b712764ac55cd59c61f314d6 WatchSource:0}: Error finding container 4192088e8b50de0b64d86c28705e817f1640d388b712764ac55cd59c61f314d6: Status 404 returned error can't find the container with id 4192088e8b50de0b64d86c28705e817f1640d388b712764ac55cd59c61f314d6 Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.910306 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" event={"ID":"cc4dd1dc-155a-4278-8382-cbe5364f80eb","Type":"ContainerStarted","Data":"94abff3472d1e0607b63bc31c0299f835e11f919cd1519eb4ad14d9b895aa57e"} Dec 03 08:40:21 crc kubenswrapper[4573]: I1203 08:40:21.967800 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:21 crc kubenswrapper[4573]: E1203 08:40:21.983412 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.483391181 +0000 UTC m=+143.051770440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.070211 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.075572 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.575546234 +0000 UTC m=+143.143925493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.080397 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.084133 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.584117768 +0000 UTC m=+143.152497027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.188215 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.188538 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.688523492 +0000 UTC m=+143.256902751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.232071 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs"] Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.302393 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.302662 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.80265158 +0000 UTC m=+143.371030829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.408594 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.409008 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:22.908991364 +0000 UTC m=+143.477370623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.417325 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-dpwlm" podStartSLOduration=123.417300862 podStartE2EDuration="2m3.417300862s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:22.413233396 +0000 UTC m=+142.981612655" watchObservedRunningTime="2025-12-03 08:40:22.417300862 +0000 UTC m=+142.985680121" Dec 03 08:40:22 crc kubenswrapper[4573]: W1203 08:40:22.439942 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c10cae9_fe93_4134_ae2f_7e242201dc5f.slice/crio-8629d95bf35528b1b982bbca877333d0cf69debb557ebcd562a2378a3f5c4991 WatchSource:0}: Error finding container 8629d95bf35528b1b982bbca877333d0cf69debb557ebcd562a2378a3f5c4991: Status 404 returned error can't find the container with id 8629d95bf35528b1b982bbca877333d0cf69debb557ebcd562a2378a3f5c4991 Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.491144 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b92rg"] Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.514550 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.514868 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.014855866 +0000 UTC m=+143.583235125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.616431 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.616575 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.116554508 +0000 UTC m=+143.684933777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.616873 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.617356 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.11734555 +0000 UTC m=+143.685724809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.727548 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.728023 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.227984406 +0000 UTC m=+143.796363665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.770333 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" podStartSLOduration=122.770312415 podStartE2EDuration="2m2.770312415s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:22.739734994 +0000 UTC m=+143.308114263" watchObservedRunningTime="2025-12-03 08:40:22.770312415 +0000 UTC m=+143.338691674" Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.782241 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94"] Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.794137 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-vphhd" podStartSLOduration=123.794117108 podStartE2EDuration="2m3.794117108s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:22.791765176 +0000 UTC m=+143.360144435" watchObservedRunningTime="2025-12-03 08:40:22.794117108 +0000 UTC m=+143.362496387" Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.813534 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j"] Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.850992 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.851367 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.351351116 +0000 UTC m=+143.919730375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.876598 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg"] Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.952977 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:22 crc kubenswrapper[4573]: E1203 08:40:22.953299 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.453283005 +0000 UTC m=+144.021662264 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:22 crc kubenswrapper[4573]: I1203 08:40:22.991737 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" event={"ID":"7807daee-9b44-4ebd-aab9-afaf91fb9746","Type":"ContainerStarted","Data":"cfd37d43304e76ce43a9b099b1cdfda31ff5ac7a1d59ba668bf76e6980da0136"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.020359 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b92rg" event={"ID":"b3c8277f-eef7-4121-8b4e-ed0df8deda4b","Type":"ContainerStarted","Data":"7176b96fb3e367c8507e309281e66fa48472f295b926e252486d657c4aec93f2"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.024268 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" event={"ID":"284d9db2-04e5-4c39-b4cf-6f001de356c4","Type":"ContainerStarted","Data":"0e37ddb0ae7b0e975fbb132d289494e7aa346bd3f027189674515ef64d10e6cb"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.059349 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.059700 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.559683971 +0000 UTC m=+144.128063240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.061702 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" event={"ID":"ee739812-3e92-42ff-a39d-fcea79a3164c","Type":"ContainerStarted","Data":"d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.078602 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" event={"ID":"9ba983f3-7837-46f0-b5ff-b8277252e43c","Type":"ContainerStarted","Data":"45065cb8e2a1b14d3ff7e46c7d8f68b2b4eba641171154f1c0f063d80b08d06c"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.102399 4573 generic.go:334] "Generic (PLEG): container finished" podID="5279cafe-fc2e-4037-8841-ac430e0903b6" containerID="531390684b042873b560426e9f200ee0207b14de4205253a1987a805d956cd1b" exitCode=0 Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.102527 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fg658" event={"ID":"5279cafe-fc2e-4037-8841-ac430e0903b6","Type":"ContainerDied","Data":"531390684b042873b560426e9f200ee0207b14de4205253a1987a805d956cd1b"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.124421 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd"] Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.132222 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr"] Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.162714 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.163891 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.663877789 +0000 UTC m=+144.232257048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.190982 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" event={"ID":"2c10cae9-fe93-4134-ae2f-7e242201dc5f","Type":"ContainerStarted","Data":"8629d95bf35528b1b982bbca877333d0cf69debb557ebcd562a2378a3f5c4991"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.263927 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.264276 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.764263117 +0000 UTC m=+144.332642376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.282895 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" event={"ID":"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059","Type":"ContainerStarted","Data":"d99af9c07deb6fc02da9a2418bec73f2eb100b2ee50f47e4d0ad4a58f7e9ff4a"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.295123 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2w8js" event={"ID":"cc4dd1dc-155a-4278-8382-cbe5364f80eb","Type":"ContainerStarted","Data":"7175f9a4768689a2c3b76cc331638935c47a71d59cc2e28a632c68cdc0a9bb89"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.315795 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" event={"ID":"9b31f7fa-c98a-45e3-934e-94e70d254fca","Type":"ContainerStarted","Data":"6dffb3350b18f0f7540ccfe1b7f9952877a315898786401f0bc6e0511a1430cc"} Dec 03 08:40:23 crc kubenswrapper[4573]: W1203 08:40:23.346350 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode85a90ac_b6d5_41c3_808f_eb66020ab534.slice/crio-1ba0371d1800e4d8176d10a2b30d57a1db2f104f803a282e639bfc8df6d5f8c8 WatchSource:0}: Error finding container 1ba0371d1800e4d8176d10a2b30d57a1db2f104f803a282e639bfc8df6d5f8c8: Status 404 returned error can't find the container with id 1ba0371d1800e4d8176d10a2b30d57a1db2f104f803a282e639bfc8df6d5f8c8 Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.346737 4573 generic.go:334] "Generic (PLEG): container finished" podID="8b5ee846-009f-4feb-a61b-a413aee18c4d" containerID="fc467a6ba3ded0ec4d79d51bc6d1a1076fb1c6c5917c8aef6b632be9db530976" exitCode=0 Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.346807 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" event={"ID":"8b5ee846-009f-4feb-a61b-a413aee18c4d","Type":"ContainerDied","Data":"fc467a6ba3ded0ec4d79d51bc6d1a1076fb1c6c5917c8aef6b632be9db530976"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.368228 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.368525 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.868510037 +0000 UTC m=+144.436889296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.441960 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8tpmr" event={"ID":"f5a7efaa-3253-40d5-9e1e-59482378cd8c","Type":"ContainerStarted","Data":"4192088e8b50de0b64d86c28705e817f1640d388b712764ac55cd59c61f314d6"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.463240 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" event={"ID":"2cf8e819-4c7b-41c6-984c-871a6ac87428","Type":"ContainerStarted","Data":"a27a2d3b3d44c72995cfce373fcec5d69655d2d4ac41c2c3c698ab65ffda8f57"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.470188 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.470479 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:23.970462586 +0000 UTC m=+144.538841845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.508541 4573 generic.go:334] "Generic (PLEG): container finished" podID="f1783a29-c599-40fe-a075-32732bb082db" containerID="f1f76a177a3941296c8bad8fbc8a855ce14307e2cf1656c46558eca86120fff6" exitCode=0 Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.508640 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" event={"ID":"f1783a29-c599-40fe-a075-32732bb082db","Type":"ContainerDied","Data":"f1f76a177a3941296c8bad8fbc8a855ce14307e2cf1656c46558eca86120fff6"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.529486 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" podStartSLOduration=124.529465361 podStartE2EDuration="2m4.529465361s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:23.529354548 +0000 UTC m=+144.097733817" watchObservedRunningTime="2025-12-03 08:40:23.529465361 +0000 UTC m=+144.097844630" Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.530290 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" event={"ID":"fef5501f-3c10-40c0-8f37-0383d15eb68f","Type":"ContainerStarted","Data":"b420c18d6d44c30fa7a7e62dc74f72a7330d26534a111d0a9593e4058994982e"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.544553 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" podStartSLOduration=122.544531126 podStartE2EDuration="2m2.544531126s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:23.463488844 +0000 UTC m=+144.031868103" watchObservedRunningTime="2025-12-03 08:40:23.544531126 +0000 UTC m=+144.112910385" Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.546666 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4twfh" event={"ID":"54b30813-ea73-4e52-b6ac-c2d1eda3da53","Type":"ContainerStarted","Data":"7092835e4ba8c9a78d8debeb60f7b16f320e8d0258f5337e2449a2417461cc0f"} Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.572539 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.572859 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.072844937 +0000 UTC m=+144.641224196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.587979 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.677993 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.680492 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.180478395 +0000 UTC m=+144.748857654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.745878 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gf25" podStartSLOduration=124.745859276 podStartE2EDuration="2m4.745859276s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:23.725235207 +0000 UTC m=+144.293614466" watchObservedRunningTime="2025-12-03 08:40:23.745859276 +0000 UTC m=+144.314238535" Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.762236 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k2b4r"] Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.814882 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.815590 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.315574042 +0000 UTC m=+144.883953301 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.846246 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bjl65"] Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.878467 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-8tpmr" podStartSLOduration=123.878448929 podStartE2EDuration="2m3.878448929s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:23.877178805 +0000 UTC m=+144.445558064" watchObservedRunningTime="2025-12-03 08:40:23.878448929 +0000 UTC m=+144.446828188" Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.891634 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.922256 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:23 crc kubenswrapper[4573]: E1203 08:40:23.922554 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.422542372 +0000 UTC m=+144.990921631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:23 crc kubenswrapper[4573]: I1203 08:40:23.951023 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-pfjsx" podStartSLOduration=124.951003608 podStartE2EDuration="2m4.951003608s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:23.947342452 +0000 UTC m=+144.515721711" watchObservedRunningTime="2025-12-03 08:40:23.951003608 +0000 UTC m=+144.519382857" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.023116 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.023408 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.523394593 +0000 UTC m=+145.091773852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.024404 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-4twfh" podStartSLOduration=6.024385039 podStartE2EDuration="6.024385039s" podCreationTimestamp="2025-12-03 08:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:24.020549128 +0000 UTC m=+144.588928387" watchObservedRunningTime="2025-12-03 08:40:24.024385039 +0000 UTC m=+144.592764298" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.124809 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.125397 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.625384924 +0000 UTC m=+145.193764183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.228737 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.230780 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.730737551 +0000 UTC m=+145.299116810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.237371 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pxklw"] Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.281011 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.281487 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.281540 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.304107 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf"] Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.331756 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.332129 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.832113907 +0000 UTC m=+145.400493166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.401238 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-smk4x"] Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.436556 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.436880 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:24.936864909 +0000 UTC m=+145.505244168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.540782 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.541222 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.041208861 +0000 UTC m=+145.609588120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.547493 4573 patch_prober.go:28] interesting pod/console-operator-58897d9998-vphhd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.547546 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-vphhd" podUID="3e8a1377-41c2-4b26-8bd1-918fdde3d19a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.561496 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km"] Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.635010 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" event={"ID":"715be642-2503-4da6-9caf-05ff95d26fc3","Type":"ContainerStarted","Data":"401ad76fb04f46c2cb4a605ce9a25911f18c20c8ba5174e90ae918b6c8950571"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.643212 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.645495 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.145474491 +0000 UTC m=+145.713853770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: W1203 08:40:24.699505 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c9346bd_44d0_47ea_9d16_594259b8a867.slice/crio-2f01d95a5ab1798092d392e1d965716ca0846ed201b9d0d2c6547ef6413f16ba WatchSource:0}: Error finding container 2f01d95a5ab1798092d392e1d965716ca0846ed201b9d0d2c6547ef6413f16ba: Status 404 returned error can't find the container with id 2f01d95a5ab1798092d392e1d965716ca0846ed201b9d0d2c6547ef6413f16ba Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.700251 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" event={"ID":"a44990ef-b3ba-4db7-98b2-7c6100a3acae","Type":"ContainerStarted","Data":"dd1f770decce4b87b293032bf73943687495835c3053f4d72e1f0c2ee6ded7c2"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.701730 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-smk4x" event={"ID":"db0d1452-dca7-4745-9f61-6ae4ae162a0f","Type":"ContainerStarted","Data":"0ac7a8cbccfd6dd5d32753efca24c2c305e90ee61e46487edd91a6dea45b8ce6"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.704826 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" event={"ID":"8b5ee846-009f-4feb-a61b-a413aee18c4d","Type":"ContainerStarted","Data":"42edd96b0d299ee4f29c2ca68e5decca8be0a47272ceebbc94a0d2dd1d2fac4c"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.708272 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.732302 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" event={"ID":"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb","Type":"ContainerStarted","Data":"9b7b010ceba815fe1cad9a9321eb5d3704864f224cda2a3338b6b93b76da3c23"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.747042 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.747448 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.247433021 +0000 UTC m=+145.815812280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.748591 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" event={"ID":"fef5501f-3c10-40c0-8f37-0383d15eb68f","Type":"ContainerStarted","Data":"a66c046c3e3b4e01bc55b0b14a7b029cfab6427cf7bef8db59c94f83e0d4d869"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.777148 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" event={"ID":"9b31f7fa-c98a-45e3-934e-94e70d254fca","Type":"ContainerStarted","Data":"3da6b4ad8391b73d449e5c14f538d4facd9a4bcf69eda0d52f574adea305eeb9"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.787349 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8tpmr" event={"ID":"f5a7efaa-3253-40d5-9e1e-59482378cd8c","Type":"ContainerStarted","Data":"2026d30e6467f42681096111ad40085f3b1158a4b62bacea9f48a77e271f8e98"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.789581 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" podStartSLOduration=125.789563604 podStartE2EDuration="2m5.789563604s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:24.776785929 +0000 UTC m=+145.345165208" watchObservedRunningTime="2025-12-03 08:40:24.789563604 +0000 UTC m=+145.357942863" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.790297 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh"] Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.827763 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" event={"ID":"9ba983f3-7837-46f0-b5ff-b8277252e43c","Type":"ContainerStarted","Data":"0c332b06a25a1758f7b5bd6459831f0c0d9e0c8eaec59ce3eb87dc215a282754"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.833200 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vzjzb" podStartSLOduration=125.833184686 podStartE2EDuration="2m5.833184686s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:24.832709853 +0000 UTC m=+145.401089132" watchObservedRunningTime="2025-12-03 08:40:24.833184686 +0000 UTC m=+145.401563945" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.843959 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b92rg" event={"ID":"b3c8277f-eef7-4121-8b4e-ed0df8deda4b","Type":"ContainerStarted","Data":"fa266013074576d456ab9ddd7e4a669addbfa76e50001ca84c176a67f601102f"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.844837 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.853790 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.855524 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.355502279 +0000 UTC m=+145.923881539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.874384 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" event={"ID":"192416c5-8025-4dd5-840b-9954ce7d009b","Type":"ContainerStarted","Data":"5c3fe42bdf4fbe0f461cf4a6ec3d31ab82f940d0ca5e8c6eb7d7fea42c7dba77"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.874612 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" event={"ID":"192416c5-8025-4dd5-840b-9954ce7d009b","Type":"ContainerStarted","Data":"960ac36ffd2fa10d4564803dec08e484ab480f57e8676ed7f8aa5a4a0d92b364"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.900794 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.900854 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.903437 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" event={"ID":"76462ae4-1687-4489-b12d-0e6817753e41","Type":"ContainerStarted","Data":"c3d75aba8a7fd69931d48706b63cbe03b2d7545daf9e82329d180c503ab7923a"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.903495 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" event={"ID":"76462ae4-1687-4489-b12d-0e6817753e41","Type":"ContainerStarted","Data":"a4a1f7dc3887ad52f29b9ccf8f2845dc3d114c182ed23c9c0306a08f4a74f9a0"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.938650 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" event={"ID":"2c10cae9-fe93-4134-ae2f-7e242201dc5f","Type":"ContainerStarted","Data":"1bc7ef0526e408e034ba5fcf74b11310b9a35a4dc23966e1e088c96dd51881a8"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.940810 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" event={"ID":"43fd56dc-b8c0-4d05-8ca2-9314ce4e6059","Type":"ContainerStarted","Data":"71b981fec9331277fa99c3f6bea463aae56851f88a08b87b03cf2b0c4822b32d"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.941480 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.955009 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sjs5t" podStartSLOduration=125.954994185 podStartE2EDuration="2m5.954994185s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:24.877843424 +0000 UTC m=+145.446222683" watchObservedRunningTime="2025-12-03 08:40:24.954994185 +0000 UTC m=+145.523373434" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.961214 4573 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-z2gvw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.961268 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" podUID="43fd56dc-b8c0-4d05-8ca2-9314ce4e6059" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.977608 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" event={"ID":"ac517135-70a0-4f3d-b19c-9db51f67b2fc","Type":"ContainerStarted","Data":"606ec1882180f318038881489984be06892d70408d63eefa00c3293654a6a0e9"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.977655 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" event={"ID":"ac517135-70a0-4f3d-b19c-9db51f67b2fc","Type":"ContainerStarted","Data":"b39be93dcabe790ebc33caeaa7a99b195de9e4a05b26179cd2ce1cf1526891c0"} Dec 03 08:40:24 crc kubenswrapper[4573]: I1203 08:40:24.978859 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:24 crc kubenswrapper[4573]: E1203 08:40:24.979180 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.479165228 +0000 UTC m=+146.047544487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.041844 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-b92rg" podStartSLOduration=126.041821878 podStartE2EDuration="2m6.041821878s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.036481178 +0000 UTC m=+145.604860457" watchObservedRunningTime="2025-12-03 08:40:25.041821878 +0000 UTC m=+145.610201137" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.041943 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7mk94" podStartSLOduration=125.041938941 podStartE2EDuration="2m5.041938941s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:24.956102223 +0000 UTC m=+145.524481482" watchObservedRunningTime="2025-12-03 08:40:25.041938941 +0000 UTC m=+145.610318200" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.067197 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" event={"ID":"2cf8e819-4c7b-41c6-984c-871a6ac87428","Type":"ContainerStarted","Data":"746c799e858274da8a76a3eb994ab6f01eef94ca3668977ce7485b73ec7f3f21"} Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.067700 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.084647 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.085853 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.585833891 +0000 UTC m=+146.154213160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.097981 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" event={"ID":"76718a7d-3779-4ff8-a60d-5bef0372d0a2","Type":"ContainerStarted","Data":"492ae7ce08efabcb2fce96316001c759c70475963fd56eac11edf186a026660c"} Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.129653 4573 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-m7kqr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" start-of-body= Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.129695 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" podUID="2cf8e819-4c7b-41c6-984c-871a6ac87428" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.137928 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-d8w5j" podStartSLOduration=125.137910104 podStartE2EDuration="2m5.137910104s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.120517288 +0000 UTC m=+145.688896547" watchObservedRunningTime="2025-12-03 08:40:25.137910104 +0000 UTC m=+145.706289363" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.140013 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-92rk6"] Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.150451 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bjl65" event={"ID":"d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf","Type":"ContainerStarted","Data":"f23f4abcc34e1dac70c9cfc58bc186eafa32115244bd2f8c44edd09e3e836649"} Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.236676 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.239001 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.738989447 +0000 UTC m=+146.307368706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.284818 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" podStartSLOduration=126.284803779 podStartE2EDuration="2m6.284803779s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.28255871 +0000 UTC m=+145.850937969" watchObservedRunningTime="2025-12-03 08:40:25.284803779 +0000 UTC m=+145.853183038" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.318492 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:25 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:25 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:25 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.318983 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.319903 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" event={"ID":"c7175f33-8e75-469b-a085-20a7553e2b27","Type":"ContainerStarted","Data":"a615abc81bccf115de70c718dd83383574ad9c3322f660c58dcb0c795dd58c42"} Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.341866 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.357687 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.857660748 +0000 UTC m=+146.426040007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.425827 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" event={"ID":"284d9db2-04e5-4c39-b4cf-6f001de356c4","Type":"ContainerStarted","Data":"111faf58961e7e3ec25c8ed22cdab06b73248dc2a290ec4c2b62bbea3741b9ef"} Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.434496 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp"] Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.457862 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.458415 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:25.958404046 +0000 UTC m=+146.526783305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.458938 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2rwtg" podStartSLOduration=125.458917749 podStartE2EDuration="2m5.458917749s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.44748464 +0000 UTC m=+146.015863899" watchObservedRunningTime="2025-12-03 08:40:25.458917749 +0000 UTC m=+146.027297028" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.459325 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2ktws"] Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.464898 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" event={"ID":"e85a90ac-b6d5-41c3-808f-eb66020ab534","Type":"ContainerStarted","Data":"1ba0371d1800e4d8176d10a2b30d57a1db2f104f803a282e639bfc8df6d5f8c8"} Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.516116 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x"] Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.517023 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-jdvbk"] Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.560306 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.563632 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.063497377 +0000 UTC m=+146.631876686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.605140 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg"] Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.605964 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" podStartSLOduration=124.605943488 podStartE2EDuration="2m4.605943488s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.584150318 +0000 UTC m=+146.152529577" watchObservedRunningTime="2025-12-03 08:40:25.605943488 +0000 UTC m=+146.174322747" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.645796 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-bjl65" podStartSLOduration=7.645778932 podStartE2EDuration="7.645778932s" podCreationTimestamp="2025-12-03 08:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.645325299 +0000 UTC m=+146.213704588" watchObservedRunningTime="2025-12-03 08:40:25.645778932 +0000 UTC m=+146.214158191" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.662809 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.663709 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.16369753 +0000 UTC m=+146.732076789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.725702 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-8qfqs" podStartSLOduration=125.725683693 podStartE2EDuration="2m5.725683693s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.72364401 +0000 UTC m=+146.292023269" watchObservedRunningTime="2025-12-03 08:40:25.725683693 +0000 UTC m=+146.294062952" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.771101 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.771421 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.27140698 +0000 UTC m=+146.839786239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.816256 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl"] Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.876804 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.877245 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.377228481 +0000 UTC m=+146.945607740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.894546 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b"] Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.895400 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" podStartSLOduration=125.895386557 podStartE2EDuration="2m5.895386557s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.891908425 +0000 UTC m=+146.460287684" watchObservedRunningTime="2025-12-03 08:40:25.895386557 +0000 UTC m=+146.463765816" Dec 03 08:40:25 crc kubenswrapper[4573]: I1203 08:40:25.978028 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:25 crc kubenswrapper[4573]: E1203 08:40:25.978414 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.47839544 +0000 UTC m=+147.046774699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.093995 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.094637 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.594621393 +0000 UTC m=+147.163000662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.195720 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.196020 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.696004967 +0000 UTC m=+147.264384226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.293209 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:26 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:26 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:26 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.293260 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.296752 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.299760 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.799740513 +0000 UTC m=+147.368119782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.398226 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.398542 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:26.89852525 +0000 UTC m=+147.466904509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.506975 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.507442 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.007430621 +0000 UTC m=+147.575809880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.525699 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" event={"ID":"c7175f33-8e75-469b-a085-20a7553e2b27","Type":"ContainerStarted","Data":"8d33d13226f26c2ea842edad1c739a64849070b5ddcc3525b935d3961a871d16"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.528469 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" event={"ID":"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb","Type":"ContainerStarted","Data":"85f95a746ca84a27b624064696f388e81a5038c924fad11a2639297537072fa3"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.533939 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bjl65" event={"ID":"d6ee6b4b-b9fb-4ce4-8be7-b037c841d4cf","Type":"ContainerStarted","Data":"12b9f2011d0ec2483e1e103d9a872bde993ae5e9dee1ebd2e4401d62b2398594"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.557587 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" event={"ID":"4932282a-3036-4392-ace4-1707a321268a","Type":"ContainerStarted","Data":"4900968b862a6c11d9a6d25a5e5581fa3dbc1043aefafaef06ef630a4f659049"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.558906 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-92rk6" event={"ID":"ca368157-8579-4618-8031-a793b862bacd","Type":"ContainerStarted","Data":"0c9a22559f0bff1da8e8ccd3986cb1578ec109609d97af0d4b91bb8a981a5bbd"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.563980 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zwvkr" podStartSLOduration=126.563967701 podStartE2EDuration="2m6.563967701s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:26.561627621 +0000 UTC m=+147.130006880" watchObservedRunningTime="2025-12-03 08:40:26.563967701 +0000 UTC m=+147.132346960" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.565598 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cj8dq" podStartSLOduration=127.565591074 podStartE2EDuration="2m7.565591074s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:25.961418225 +0000 UTC m=+146.529797484" watchObservedRunningTime="2025-12-03 08:40:26.565591074 +0000 UTC m=+147.133970333" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.581871 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" event={"ID":"f1783a29-c599-40fe-a075-32732bb082db","Type":"ContainerStarted","Data":"362a2d64ee74517e0ae3e9c108d4cebf1b48e1d7ad7c0dafd830e55ffd17ae52"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.594882 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" event={"ID":"73472804-4ec7-4d1a-9a34-c6cb9e11b709","Type":"ContainerStarted","Data":"b7a2beb102821ed90642889d5f726456e8d75bce0003da4c57e13b1069d5f373"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.604626 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" event={"ID":"01a2ef90-923e-4843-91bd-5e367c232a0c","Type":"ContainerStarted","Data":"7deb48f667c7eaba0943218a97a8a32e2de06b63256caa5a107cc15041dd175a"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.609863 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.611270 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.111250899 +0000 UTC m=+147.679630168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.619960 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fg658" event={"ID":"5279cafe-fc2e-4037-8841-ac430e0903b6","Type":"ContainerStarted","Data":"0026cb41f377dc0c82ad21acd5b32974bbfb6cfbee87d9bfd8967352d64d3448"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.633198 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" event={"ID":"74479bf9-6a28-41a1-9a70-dca91bfbb855","Type":"ContainerStarted","Data":"015478e228a230922492a8ec20f9b1713d5fe3847d7a5c81bf81511749bc4b15"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.640676 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" event={"ID":"f6e10744-6f93-416a-a6aa-4a7a07c76eff","Type":"ContainerStarted","Data":"7accaac2544e969d27e833549cfc79ae2ff446046fb3ff8df43278c52d5412c3"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.652786 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" event={"ID":"9ba983f3-7837-46f0-b5ff-b8277252e43c","Type":"ContainerStarted","Data":"9107e689416befafa28f54a5ac82caf5402b8e307c9ab8ce6d95a0f8c638392e"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.657675 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" event={"ID":"6c9346bd-44d0-47ea-9d16-594259b8a867","Type":"ContainerStarted","Data":"03bdb691c050155305900452e8378c1c113e125b6eea3989719f9e49f93f781c"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.657729 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" event={"ID":"6c9346bd-44d0-47ea-9d16-594259b8a867","Type":"ContainerStarted","Data":"2f01d95a5ab1798092d392e1d965716ca0846ed201b9d0d2c6547ef6413f16ba"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.674269 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" event={"ID":"e85a90ac-b6d5-41c3-808f-eb66020ab534","Type":"ContainerStarted","Data":"5bf03c950e08a838ee8fd7592be781e2e9f54d3adce8e3face9a39c5145c579f"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.685631 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" event={"ID":"5f7280fb-cb08-4761-9d89-2f356d797c89","Type":"ContainerStarted","Data":"f3d5740624823e4817f3629b97ff11403c907038527f7cd53564869b9bc3a823"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.690363 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" event={"ID":"aef07f9d-ddd0-49a3-b1cb-7841f061a24e","Type":"ContainerStarted","Data":"2880528bd0ccf128eac7da399628e471d992a3768603c495ee8821ea5adff05d"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.697892 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" podStartSLOduration=125.697876957 podStartE2EDuration="2m5.697876957s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:26.614846674 +0000 UTC m=+147.183225953" watchObservedRunningTime="2025-12-03 08:40:26.697876957 +0000 UTC m=+147.266256216" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.698928 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pj2pz" podStartSLOduration=126.698923235 podStartE2EDuration="2m6.698923235s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:26.697562679 +0000 UTC m=+147.265941948" watchObservedRunningTime="2025-12-03 08:40:26.698923235 +0000 UTC m=+147.267302494" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.714304 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" event={"ID":"a44990ef-b3ba-4db7-98b2-7c6100a3acae","Type":"ContainerStarted","Data":"e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.714353 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.714646 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.214634436 +0000 UTC m=+147.783013695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.715018 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.718652 4573 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k2b4r container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.718717 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.723172 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" event={"ID":"fd99d655-f2f2-451b-a058-2a70e262fb2d","Type":"ContainerStarted","Data":"6693f31466f92ecef83ab3b5f62b7d541fdefcea8ea771a1e314644a048070c3"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.723239 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" event={"ID":"fd99d655-f2f2-451b-a058-2a70e262fb2d","Type":"ContainerStarted","Data":"8c401c0bc3639596166cc42e7ee0bf2c4f446f93d228a992e1540d377f3c952f"} Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.728707 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.728762 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.728806 4573 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-m7kqr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" start-of-body= Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.728823 4573 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-z2gvw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.728874 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" podUID="2cf8e819-4c7b-41c6-984c-871a6ac87428" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.728888 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" podUID="43fd56dc-b8c0-4d05-8ca2-9314ce4e6059" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.737913 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zt5bd" podStartSLOduration=126.737901205 podStartE2EDuration="2m6.737901205s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:26.733614433 +0000 UTC m=+147.301993692" watchObservedRunningTime="2025-12-03 08:40:26.737901205 +0000 UTC m=+147.306280464" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.754700 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-qdnvh" podStartSLOduration=125.754684744 podStartE2EDuration="2m5.754684744s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:26.753934325 +0000 UTC m=+147.322313584" watchObservedRunningTime="2025-12-03 08:40:26.754684744 +0000 UTC m=+147.323064013" Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.815777 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.815898 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.315872637 +0000 UTC m=+147.884251886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.820689 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.821754 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.32174188 +0000 UTC m=+147.890121139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.921480 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:26 crc kubenswrapper[4573]: E1203 08:40:26.921722 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.421678227 +0000 UTC m=+147.990057486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.943518 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:40:26 crc kubenswrapper[4573]: I1203 08:40:26.943583 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.023700 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.024120 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.524103739 +0000 UTC m=+148.092482998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.124117 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.124389 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.624359823 +0000 UTC m=+148.192739082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.124450 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.124494 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.124764 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.624752225 +0000 UTC m=+148.193131484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.148453 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.225340 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.225529 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.725503583 +0000 UTC m=+148.293882842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.225598 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.225666 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.225714 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.225768 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.226338 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.726329914 +0000 UTC m=+148.294709173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.228595 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.230633 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.268827 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.284294 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.286074 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:27 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:27 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:27 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.286109 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.326330 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.326560 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.826545578 +0000 UTC m=+148.394924827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.436361 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.437140 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:27.937122663 +0000 UTC m=+148.505501922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.535915 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.539963 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.540252 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:28.040237793 +0000 UTC m=+148.608617052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.589022 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" podStartSLOduration=127.589004999 podStartE2EDuration="2m7.589004999s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:26.77321192 +0000 UTC m=+147.341591179" watchObservedRunningTime="2025-12-03 08:40:27.589004999 +0000 UTC m=+148.157384258" Dec 03 08:40:27 crc kubenswrapper[4573]: W1203 08:40:27.602670 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-42d35674bd66cfc44ce6796ce6be5481d900303430b695e2c1add9b4e2daf9b1 WatchSource:0}: Error finding container 42d35674bd66cfc44ce6796ce6be5481d900303430b695e2c1add9b4e2daf9b1: Status 404 returned error can't find the container with id 42d35674bd66cfc44ce6796ce6be5481d900303430b695e2c1add9b4e2daf9b1 Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.641341 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.641666 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:28.141653848 +0000 UTC m=+148.710033097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.867473 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.867882 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:28.36786198 +0000 UTC m=+148.936241239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.868018 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 08:40:27 crc kubenswrapper[4573]: I1203 08:40:27.974564 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:27 crc kubenswrapper[4573]: E1203 08:40:27.975326 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:28.475311693 +0000 UTC m=+149.043690942 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.077137 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.077542 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:28.57751739 +0000 UTC m=+149.145896649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.091873 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fg658" event={"ID":"5279cafe-fc2e-4037-8841-ac430e0903b6","Type":"ContainerStarted","Data":"165fb2a391276fc282d9131850b17ef0743c98c9be4f5be3c40a8551e31130a6"} Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.128938 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" event={"ID":"6c9346bd-44d0-47ea-9d16-594259b8a867","Type":"ContainerStarted","Data":"4e6982a777387c2c93bc04f0bdc0848f29badb9beea3da48e40eff582d3f6063"} Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.135199 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" event={"ID":"715be642-2503-4da6-9caf-05ff95d26fc3","Type":"ContainerStarted","Data":"45a22614de504e564b2b1a1f0edce728c345cf9d118844fc04b8314834c46372"} Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.143023 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-fg658" podStartSLOduration=128.143001294 podStartE2EDuration="2m8.143001294s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:28.140723154 +0000 UTC m=+148.709102413" watchObservedRunningTime="2025-12-03 08:40:28.143001294 +0000 UTC m=+148.711380553" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.151326 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.154717 4573 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4rx5b container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" start-of-body= Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.154769 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" podUID="01a2ef90-923e-4843-91bd-5e367c232a0c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.179244 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.183041 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:28.683015331 +0000 UTC m=+149.251394590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.233977 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" event={"ID":"5f7280fb-cb08-4761-9d89-2f356d797c89","Type":"ContainerStarted","Data":"9f56a01284730c85df5a2c56910111ca43dfdb7918a46d73c55566f62bdb9ecc"} Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.254637 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.256775 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ea4265f0de251a96f1812007350c46ee7b9a08402eca7cc75375d523dc44bcfb"} Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.263378 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" podStartSLOduration=128.263356335 podStartE2EDuration="2m8.263356335s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:28.216232121 +0000 UTC m=+148.784611380" watchObservedRunningTime="2025-12-03 08:40:28.263356335 +0000 UTC m=+148.831735594" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.263495 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mr8km" podStartSLOduration=128.263489109 podStartE2EDuration="2m8.263489109s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:28.258657262 +0000 UTC m=+148.827036521" watchObservedRunningTime="2025-12-03 08:40:28.263489109 +0000 UTC m=+148.831868388" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.280949 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"42d35674bd66cfc44ce6796ce6be5481d900303430b695e2c1add9b4e2daf9b1"} Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.283670 4573 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k2b4r container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.283724 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.284398 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.284956 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:28.784941441 +0000 UTC m=+149.353320700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.285984 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.286021 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.289524 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:28 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:28 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:28 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.289833 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.356178 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" podStartSLOduration=127.356157425 podStartE2EDuration="2m7.356157425s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:28.343875524 +0000 UTC m=+148.912254793" watchObservedRunningTime="2025-12-03 08:40:28.356157425 +0000 UTC m=+148.924536694" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.383808 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2gvw" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.401077 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.403462 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:28.903443433 +0000 UTC m=+149.471822732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.502612 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.503329 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.003309948 +0000 UTC m=+149.571689207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.558660 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" podStartSLOduration=127.558646246 podStartE2EDuration="2m7.558646246s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:28.556502341 +0000 UTC m=+149.124881600" watchObservedRunningTime="2025-12-03 08:40:28.558646246 +0000 UTC m=+149.127025505" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.559651 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2ktws" podStartSLOduration=128.559645993 podStartE2EDuration="2m8.559645993s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:28.459529361 +0000 UTC m=+149.027908620" watchObservedRunningTime="2025-12-03 08:40:28.559645993 +0000 UTC m=+149.128025252" Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.604830 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.605416 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.105401921 +0000 UTC m=+149.673781180 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.706564 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.706916 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.206884627 +0000 UTC m=+149.775263886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.807979 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.808410 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.308396436 +0000 UTC m=+149.876775695 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.909895 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.910075 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.410024037 +0000 UTC m=+149.978403296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:28 crc kubenswrapper[4573]: I1203 08:40:28.910272 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:28 crc kubenswrapper[4573]: E1203 08:40:28.910667 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.410653203 +0000 UTC m=+149.979032462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.012463 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.012604 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.512581321 +0000 UTC m=+150.080960580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.012776 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.013166 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.513154377 +0000 UTC m=+150.081533636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.066847 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.067212 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.114334 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.114770 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.614755307 +0000 UTC m=+150.183134566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.176150 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.176186 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.192242 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-vphhd" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.219968 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.221165 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.721150502 +0000 UTC m=+150.289529761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.285332 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:29 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:29 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:29 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.285388 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.287117 4573 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-l7qrz container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.287177 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" podUID="8b5ee846-009f-4feb-a61b-a413aee18c4d" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.291871 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" event={"ID":"0c5fdb5e-0dd8-47d5-83f2-d77eb077a6bb","Type":"ContainerStarted","Data":"295687f4901be19d11fdb45369b6e7944987ec29d7c2833784c70c81da838368"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.303782 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-92rk6" event={"ID":"ca368157-8579-4618-8031-a793b862bacd","Type":"ContainerStarted","Data":"6908a683de5d455c0ebea3ed8b589f731ad685c73fe4836d98fa0f5d515e9c9f"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.303828 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-92rk6" event={"ID":"ca368157-8579-4618-8031-a793b862bacd","Type":"ContainerStarted","Data":"e90ec36134564a84974aa3bf5a0c056e048edeb1a5374115f15b5e40f83c8bc4"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.303936 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.306277 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0aa38c381aa0bdadb37fdaf5e0214c7cb5a6032ef21f33d9aab730c75232d1ab"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.306777 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.309006 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" podStartSLOduration=128.308997732 podStartE2EDuration="2m8.308997732s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:28.872295128 +0000 UTC m=+149.440674397" watchObservedRunningTime="2025-12-03 08:40:29.308997732 +0000 UTC m=+149.877376991" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.320666 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.321030 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:29.821015898 +0000 UTC m=+150.389395157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.325838 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" event={"ID":"715be642-2503-4da6-9caf-05ff95d26fc3","Type":"ContainerStarted","Data":"2b56fc74136e36d5fe578a70c5f7117548eae037d54151990671a38a43d04bd0"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.334000 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"23ca72054aa7af5bfcc6b6ac2499c98458d3a2ff94d853943f9c5efa49f818b4"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.343186 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" event={"ID":"f6e10744-6f93-416a-a6aa-4a7a07c76eff","Type":"ContainerStarted","Data":"f7c1a128200412cb6375d91505f9f917ca992398f5c00d5115de05e29d49fbca"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.344081 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.344238 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.344260 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.349664 4573 patch_prober.go:28] interesting pod/console-f9d7485db-dpwlm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.349714 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-dpwlm" podUID="bdc5518a-70d7-4f7f-bb7a-3320cb61f607" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.349669 4573 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-z2tsl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.349765 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" podUID="f6e10744-6f93-416a-a6aa-4a7a07c76eff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.387066 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-4j98x" event={"ID":"4932282a-3036-4392-ace4-1707a321268a","Type":"ContainerStarted","Data":"8e4c181aa5cfbdaaec1f1266ce9e1c4591b389f30775a12873a14d393e584237"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.399412 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" event={"ID":"aef07f9d-ddd0-49a3-b1cb-7841f061a24e","Type":"ContainerStarted","Data":"a99f588bd9217a2a93f6971fbaf75d2ed960ae5beb1a75e49f307852333db29b"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.399488 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" event={"ID":"aef07f9d-ddd0-49a3-b1cb-7841f061a24e","Type":"ContainerStarted","Data":"9ed773dc3850d87e3afc1e653595ae7559135480aa4c2fe7ab6f2f8c570ee0ae"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.403254 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-smk4x" event={"ID":"db0d1452-dca7-4745-9f61-6ae4ae162a0f","Type":"ContainerStarted","Data":"3b3bf734a4928ac9676bab3c8cdfc87e23ddf7b7bb5fe098fe38f5e4742e2a42"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.451016 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-jdvbk" event={"ID":"73472804-4ec7-4d1a-9a34-c6cb9e11b709","Type":"ContainerStarted","Data":"ab450c6ab2f127c6bf36dd65203d4c253918aadf9f2178d14b9cc1f1ade223ed"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.459222 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" event={"ID":"01a2ef90-923e-4843-91bd-5e367c232a0c","Type":"ContainerStarted","Data":"b7b1117721d715de3de45980ee869eebb2f97d1936fd12c25b9a6cbd8071f803"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.460211 4573 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4rx5b container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" start-of-body= Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.460245 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" podUID="01a2ef90-923e-4843-91bd-5e367c232a0c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.476418 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" event={"ID":"74479bf9-6a28-41a1-9a70-dca91bfbb855","Type":"ContainerStarted","Data":"f0ba9d8849a5615e60a927ebaf111b43d96fb329869f4383eb39dc0a87d9233b"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.476466 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" event={"ID":"74479bf9-6a28-41a1-9a70-dca91bfbb855","Type":"ContainerStarted","Data":"fdb792bc620f430ba614bbe8f2502f66fb21d655f6364ddf1c50305d346d95f5"} Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.478369 4573 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k2b4r container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.478403 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.504556 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.506648 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.006633687 +0000 UTC m=+150.575012946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.608122 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.608523 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.108505414 +0000 UTC m=+150.676884673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.621637 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.725964 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.726314 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.226301338 +0000 UTC m=+150.794680597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.726889 4573 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-l7qrz container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.726938 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" podUID="8b5ee846-009f-4feb-a61b-a413aee18c4d" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.824597 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-7lnsg" podStartSLOduration=129.824572401 podStartE2EDuration="2m9.824572401s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:29.698975462 +0000 UTC m=+150.267354721" watchObservedRunningTime="2025-12-03 08:40:29.824572401 +0000 UTC m=+150.392951660" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.825416 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" podStartSLOduration=128.825409923 podStartE2EDuration="2m8.825409923s" podCreationTimestamp="2025-12-03 08:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:29.82225585 +0000 UTC m=+150.390635109" watchObservedRunningTime="2025-12-03 08:40:29.825409923 +0000 UTC m=+150.393789182" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.826689 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.827027 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.327010955 +0000 UTC m=+150.895390214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.842160 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.865458 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wd2xp" podStartSLOduration=129.865443531 podStartE2EDuration="2m9.865443531s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:29.86273064 +0000 UTC m=+150.431109899" watchObservedRunningTime="2025-12-03 08:40:29.865443531 +0000 UTC m=+150.433822790" Dec 03 08:40:29 crc kubenswrapper[4573]: I1203 08:40:29.996921 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:29 crc kubenswrapper[4573]: E1203 08:40:29.997316 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.497303154 +0000 UTC m=+151.065682413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.061425 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-92rk6" podStartSLOduration=12.061405892 podStartE2EDuration="12.061405892s" podCreationTimestamp="2025-12-03 08:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:29.996513423 +0000 UTC m=+150.564892682" watchObservedRunningTime="2025-12-03 08:40:30.061405892 +0000 UTC m=+150.629785151" Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.062796 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-pxklw" podStartSLOduration=130.062788818 podStartE2EDuration="2m10.062788818s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:30.059761968 +0000 UTC m=+150.628141227" watchObservedRunningTime="2025-12-03 08:40:30.062788818 +0000 UTC m=+150.631168067" Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.098025 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.098368 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.598353659 +0000 UTC m=+151.166732918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.200461 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.200770 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.700755841 +0000 UTC m=+151.269135100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.287209 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:30 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:30 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:30 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.287271 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.387402 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.387937 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.88791406 +0000 UTC m=+151.456293319 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.489174 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.489704 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:30.989683235 +0000 UTC m=+151.558062564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.492632 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"bbe4955267983aaf73664cf595dcbaacf22b2df807bcbba4d08ee954d43f61f1"} Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.497519 4573 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-z2tsl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.497560 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" podUID="f6e10744-6f93-416a-a6aa-4a7a07c76eff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.497936 4573 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4rx5b container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" start-of-body= Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.497970 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" podUID="01a2ef90-923e-4843-91bd-5e367c232a0c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.513276 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8k9cq" Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.590852 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.591186 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.091161692 +0000 UTC m=+151.659540951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.591496 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.596385 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.096368228 +0000 UTC m=+151.664747497 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.695826 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.695969 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.195938835 +0000 UTC m=+151.764318094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.696218 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.696513 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.1965017 +0000 UTC m=+151.764880959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.770772 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.807877 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.809095 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.309072558 +0000 UTC m=+151.877451817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.909860 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:30 crc kubenswrapper[4573]: E1203 08:40:30.910355 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.410342559 +0000 UTC m=+151.978721818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:30 crc kubenswrapper[4573]: I1203 08:40:30.920976 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.011390 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.012513 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.512497753 +0000 UTC m=+152.080877012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.157135 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.157570 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.657555701 +0000 UTC m=+152.225934960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.162419 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.162478 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.162412 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.162562 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.258114 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.258487 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.758471004 +0000 UTC m=+152.326850263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.279621 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.287067 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:31 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:31 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:31 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.287124 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.314381 4573 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k2b4r container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.314448 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.314402 4573 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k2b4r container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.314892 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.370271 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.370586 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:31.870573099 +0000 UTC m=+152.438952358 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.516522 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.516846 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.016832028 +0000 UTC m=+152.585211277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.548007 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-smk4x" event={"ID":"db0d1452-dca7-4745-9f61-6ae4ae162a0f","Type":"ContainerStarted","Data":"1b4ce53b3177de2c488dc85ef3016fc9d1e821d41c8b634ff131c2e4503f6ba0"} Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.554458 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"98015d94009eb52982337082baf72ef49528448787819a37813adc2846ec0417"} Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.555222 4573 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-z2tsl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.555273 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" podUID="f6e10744-6f93-416a-a6aa-4a7a07c76eff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.612675 4573 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-z2tsl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.612717 4573 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-z2tsl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.612749 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" podUID="f6e10744-6f93-416a-a6aa-4a7a07c76eff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.612765 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" podUID="f6e10744-6f93-416a-a6aa-4a7a07c76eff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.617635 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.617981 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.117969736 +0000 UTC m=+152.686348995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.695402 4573 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4rx5b container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.695475 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" podUID="01a2ef90-923e-4843-91bd-5e367c232a0c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.695987 4573 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4rx5b container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" start-of-body= Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.696023 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" podUID="01a2ef90-923e-4843-91bd-5e367c232a0c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.719615 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.720699 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.220683815 +0000 UTC m=+152.789063074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.822010 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.822399 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.322386938 +0000 UTC m=+152.890766197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.923010 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.923169 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.423143226 +0000 UTC m=+152.991522485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:31 crc kubenswrapper[4573]: I1203 08:40:31.923533 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:31 crc kubenswrapper[4573]: E1203 08:40:31.923867 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.423836835 +0000 UTC m=+152.992216094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.025036 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.025321 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.525304931 +0000 UTC m=+153.093684190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.125888 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.126233 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.626222334 +0000 UTC m=+153.194601593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.227092 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.227481 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.727466334 +0000 UTC m=+153.295845593 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.282739 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:32 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:32 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:32 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.282798 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.333949 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.334344 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.834328822 +0000 UTC m=+153.402708081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.434719 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.434877 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.934855664 +0000 UTC m=+153.503234933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.434973 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.435273 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:32.935264625 +0000 UTC m=+153.503643884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.536499 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.536744 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.036709621 +0000 UTC m=+153.605088880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.569441 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-smk4x" event={"ID":"db0d1452-dca7-4745-9f61-6ae4ae162a0f","Type":"ContainerStarted","Data":"744b6834ef77b4d7a51c9897c47887bcf001255f329c5e1116135708912ed8fd"} Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.638077 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.638525 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.138511047 +0000 UTC m=+153.706890306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.738702 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.738895 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.238870164 +0000 UTC m=+153.807249423 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.739173 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.739465 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.239453579 +0000 UTC m=+153.807832838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.849092 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.849705 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.349681475 +0000 UTC m=+153.918060734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.951816 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:32 crc kubenswrapper[4573]: E1203 08:40:32.952280 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.452268311 +0000 UTC m=+154.020647570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.985144 4573 patch_prober.go:28] interesting pod/apiserver-76f77b778f-fg658 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]log ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]etcd ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/max-in-flight-filter ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 08:40:32 crc kubenswrapper[4573]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 08:40:32 crc kubenswrapper[4573]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/openshift.io-startinformers ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 08:40:32 crc kubenswrapper[4573]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 08:40:32 crc kubenswrapper[4573]: livez check failed Dec 03 08:40:32 crc kubenswrapper[4573]: I1203 08:40:32.985202 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-fg658" podUID="5279cafe-fc2e-4037-8841-ac430e0903b6" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.053534 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:33 crc kubenswrapper[4573]: E1203 08:40:33.053759 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.553740358 +0000 UTC m=+154.122119617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.054130 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:33 crc kubenswrapper[4573]: E1203 08:40:33.055727 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.55571315 +0000 UTC m=+154.124092409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.155059 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:33 crc kubenswrapper[4573]: E1203 08:40:33.155470 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.655452921 +0000 UTC m=+154.223832180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.178398 4573 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.250448 4573 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T08:40:33.178428492Z","Handler":null,"Name":""} Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.256442 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:33 crc kubenswrapper[4573]: E1203 08:40:33.256816 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 08:40:33.756801324 +0000 UTC m=+154.325180583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-h7gd4" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.266741 4573 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.266784 4573 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.283706 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:33 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:33 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:33 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.283764 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.357200 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.371834 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.372436 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.375266 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.376288 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.470711 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.508839 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.560510 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2d9abac-8816-4343-979e-805a0d20095e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d2d9abac-8816-4343-979e-805a0d20095e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.560874 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.560943 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2d9abac-8816-4343-979e-805a0d20095e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d2d9abac-8816-4343-979e-805a0d20095e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.563254 4573 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.563288 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.576361 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-smk4x" event={"ID":"db0d1452-dca7-4745-9f61-6ae4ae162a0f","Type":"ContainerStarted","Data":"28e13b5cc02527c6aba733efa422ed43cdba3347433141e7a4ab98bb735a29ad"} Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.601518 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-smk4x" podStartSLOduration=15.60149833 podStartE2EDuration="15.60149833s" podCreationTimestamp="2025-12-03 08:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:33.601019397 +0000 UTC m=+154.169398646" watchObservedRunningTime="2025-12-03 08:40:33.60149833 +0000 UTC m=+154.169877589" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.752363 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2d9abac-8816-4343-979e-805a0d20095e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d2d9abac-8816-4343-979e-805a0d20095e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.752403 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2d9abac-8816-4343-979e-805a0d20095e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d2d9abac-8816-4343-979e-805a0d20095e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.753310 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2d9abac-8816-4343-979e-805a0d20095e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d2d9abac-8816-4343-979e-805a0d20095e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.778530 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2d9abac-8816-4343-979e-805a0d20095e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d2d9abac-8816-4343-979e-805a0d20095e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.821113 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-h7gd4\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:33 crc kubenswrapper[4573]: I1203 08:40:33.986859 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.109326 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.112897 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.113591 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5bdk6"] Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.115651 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.130140 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.138965 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5bdk6"] Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.193222 4573 patch_prober.go:28] interesting pod/apiserver-76f77b778f-fg658 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]log ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]etcd ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/max-in-flight-filter ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 08:40:34 crc kubenswrapper[4573]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/openshift.io-startinformers ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 08:40:34 crc kubenswrapper[4573]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 08:40:34 crc kubenswrapper[4573]: livez check failed Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.193628 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-fg658" podUID="5279cafe-fc2e-4037-8841-ac430e0903b6" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.263728 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnqgr\" (UniqueName: \"kubernetes.io/projected/bdf17da2-3e3a-4940-ae7c-680e6af80d09-kube-api-access-pnqgr\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.263779 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-utilities\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.263827 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-catalog-content\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.291308 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:34 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:34 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:34 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.291380 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.295019 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-722wl"] Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.295913 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.317798 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.334419 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-722wl"] Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.367623 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnqgr\" (UniqueName: \"kubernetes.io/projected/bdf17da2-3e3a-4940-ae7c-680e6af80d09-kube-api-access-pnqgr\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.367693 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-utilities\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.367754 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-catalog-content\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.368442 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-catalog-content\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.369016 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-utilities\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.420304 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnqgr\" (UniqueName: \"kubernetes.io/projected/bdf17da2-3e3a-4940-ae7c-680e6af80d09-kube-api-access-pnqgr\") pod \"certified-operators-5bdk6\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.465768 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.469572 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-utilities\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.469641 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-catalog-content\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.469667 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95rk5\" (UniqueName: \"kubernetes.io/projected/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-kube-api-access-95rk5\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.550278 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mbtqh"] Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.551454 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.574148 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95rk5\" (UniqueName: \"kubernetes.io/projected/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-kube-api-access-95rk5\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.574286 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-utilities\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.574356 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-catalog-content\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.574997 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-catalog-content\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.600498 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-utilities\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.615819 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbtqh"] Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.640655 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95rk5\" (UniqueName: \"kubernetes.io/projected/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-kube-api-access-95rk5\") pod \"community-operators-722wl\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.676994 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-catalog-content\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.677036 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-utilities\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.677084 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmjd9\" (UniqueName: \"kubernetes.io/projected/93486cc3-d649-430e-8297-723a34105ef9-kube-api-access-wmjd9\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.702462 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-722wl" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.721512 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f77fx"] Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.726080 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.773437 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f77fx"] Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.782601 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmjd9\" (UniqueName: \"kubernetes.io/projected/93486cc3-d649-430e-8297-723a34105ef9-kube-api-access-wmjd9\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.783033 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-catalog-content\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.783296 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-utilities\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.784875 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-catalog-content\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.785215 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-utilities\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.848985 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmjd9\" (UniqueName: \"kubernetes.io/projected/93486cc3-d649-430e-8297-723a34105ef9-kube-api-access-wmjd9\") pod \"certified-operators-mbtqh\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.880313 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.900571 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-catalog-content\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.900613 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-utilities\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:34 crc kubenswrapper[4573]: I1203 08:40:34.900649 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbwkb\" (UniqueName: \"kubernetes.io/projected/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-kube-api-access-nbwkb\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.002752 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-catalog-content\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.002828 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-utilities\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.002871 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbwkb\" (UniqueName: \"kubernetes.io/projected/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-kube-api-access-nbwkb\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.004256 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-utilities\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.004410 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-catalog-content\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.039811 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbwkb\" (UniqueName: \"kubernetes.io/projected/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-kube-api-access-nbwkb\") pod \"community-operators-f77fx\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.053421 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.122158 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.286583 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h7gd4"] Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.312028 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:35 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:35 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:35 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.312091 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.340555 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5bdk6"] Dec 03 08:40:35 crc kubenswrapper[4573]: W1203 08:40:35.406847 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdf17da2_3e3a_4940_ae7c_680e6af80d09.slice/crio-631f98eb8469838fb3423ce4aa54d8b2c07c736f28c0c0b668d8fbd8fda1dd8a WatchSource:0}: Error finding container 631f98eb8469838fb3423ce4aa54d8b2c07c736f28c0c0b668d8fbd8fda1dd8a: Status 404 returned error can't find the container with id 631f98eb8469838fb3423ce4aa54d8b2c07c736f28c0c0b668d8fbd8fda1dd8a Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.568873 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f77fx"] Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.666829 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" event={"ID":"468b20da-e9e1-43e5-a265-54d131b1b66a","Type":"ContainerStarted","Data":"47c2d81e0abf8d42085fe6c7147c17319d2e641aa99cb7405246348ee630bf39"} Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.669077 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bdk6" event={"ID":"bdf17da2-3e3a-4940-ae7c-680e6af80d09","Type":"ContainerStarted","Data":"631f98eb8469838fb3423ce4aa54d8b2c07c736f28c0c0b668d8fbd8fda1dd8a"} Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.689730 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f77fx" event={"ID":"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9","Type":"ContainerStarted","Data":"e99d8d8f966904eef2140be92db550a377cc5d704f0194ae0b2622f4cf3f2ba9"} Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.694932 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d2d9abac-8816-4343-979e-805a0d20095e","Type":"ContainerStarted","Data":"b54c8625aefaedfde1d63f2882c2ac67434af3e70686954141f94584fc4fa558"} Dec 03 08:40:35 crc kubenswrapper[4573]: I1203 08:40:35.714077 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-722wl"] Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.048815 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbtqh"] Dec 03 08:40:36 crc kubenswrapper[4573]: W1203 08:40:36.059270 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93486cc3_d649_430e_8297_723a34105ef9.slice/crio-370b20d43851103106d464c74631d8c3d3016edd72ecf2e46a3b54553503d80c WatchSource:0}: Error finding container 370b20d43851103106d464c74631d8c3d3016edd72ecf2e46a3b54553503d80c: Status 404 returned error can't find the container with id 370b20d43851103106d464c74631d8c3d3016edd72ecf2e46a3b54553503d80c Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.090959 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fzg9q"] Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.092626 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.096901 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.118594 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fzg9q"] Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.140837 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.142216 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.149351 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.151473 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.162059 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.365379 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.365485 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-utilities\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.365579 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:36 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:36 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:36 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.365656 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.365676 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-catalog-content\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.365993 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.366073 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2g6r\" (UniqueName: \"kubernetes.io/projected/91e26815-0b3b-4855-be0a-603e52a2d02e-kube-api-access-t2g6r\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.467901 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.467947 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2g6r\" (UniqueName: \"kubernetes.io/projected/91e26815-0b3b-4855-be0a-603e52a2d02e-kube-api-access-t2g6r\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.467987 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.468015 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-utilities\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.468079 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-catalog-content\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.468073 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.468539 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-utilities\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.468676 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-catalog-content\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.474412 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kndk9"] Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.476351 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.493022 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kndk9"] Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.500084 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2g6r\" (UniqueName: \"kubernetes.io/projected/91e26815-0b3b-4855-be0a-603e52a2d02e-kube-api-access-t2g6r\") pod \"redhat-marketplace-fzg9q\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.504030 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.571649 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-catalog-content\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.571733 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-utilities\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.571763 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdfwb\" (UniqueName: \"kubernetes.io/projected/5b352afe-f1f8-4b25-b6a0-3ec01a374536-kube-api-access-qdfwb\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.673184 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-utilities\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.673323 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdfwb\" (UniqueName: \"kubernetes.io/projected/5b352afe-f1f8-4b25-b6a0-3ec01a374536-kube-api-access-qdfwb\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.673498 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-catalog-content\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.674195 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-catalog-content\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.674280 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-utilities\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.694113 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdfwb\" (UniqueName: \"kubernetes.io/projected/5b352afe-f1f8-4b25-b6a0-3ec01a374536-kube-api-access-qdfwb\") pod \"redhat-marketplace-kndk9\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.702230 4573 generic.go:334] "Generic (PLEG): container finished" podID="93486cc3-d649-430e-8297-723a34105ef9" containerID="ccdc12851d76fd4700cc00c4270ab9dd3717161b09ad4e8c0b984ca0dd2b5409" exitCode=0 Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.702338 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbtqh" event={"ID":"93486cc3-d649-430e-8297-723a34105ef9","Type":"ContainerDied","Data":"ccdc12851d76fd4700cc00c4270ab9dd3717161b09ad4e8c0b984ca0dd2b5409"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.702381 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbtqh" event={"ID":"93486cc3-d649-430e-8297-723a34105ef9","Type":"ContainerStarted","Data":"370b20d43851103106d464c74631d8c3d3016edd72ecf2e46a3b54553503d80c"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.704519 4573 generic.go:334] "Generic (PLEG): container finished" podID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerID="d409692cda2225631b89b257e2ebbea2dbcfbfe2aa714fc2cec8b93eb5e1b121" exitCode=0 Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.704602 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-722wl" event={"ID":"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf","Type":"ContainerDied","Data":"d409692cda2225631b89b257e2ebbea2dbcfbfe2aa714fc2cec8b93eb5e1b121"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.704715 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-722wl" event={"ID":"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf","Type":"ContainerStarted","Data":"5db4a32a165f7936b41e1e31ea0ee87c89e40b64400940151e908bafb8abcc61"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.711749 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.723029 4573 generic.go:334] "Generic (PLEG): container finished" podID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerID="b87ce9be9ba90ab57297935dfbc1bf5c9e17f8fae263242316520a725cf00d98" exitCode=0 Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.723136 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f77fx" event={"ID":"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9","Type":"ContainerDied","Data":"b87ce9be9ba90ab57297935dfbc1bf5c9e17f8fae263242316520a725cf00d98"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.738581 4573 generic.go:334] "Generic (PLEG): container finished" podID="2c10cae9-fe93-4134-ae2f-7e242201dc5f" containerID="1bc7ef0526e408e034ba5fcf74b11310b9a35a4dc23966e1e088c96dd51881a8" exitCode=0 Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.738660 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" event={"ID":"2c10cae9-fe93-4134-ae2f-7e242201dc5f","Type":"ContainerDied","Data":"1bc7ef0526e408e034ba5fcf74b11310b9a35a4dc23966e1e088c96dd51881a8"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.746128 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.781359 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d2d9abac-8816-4343-979e-805a0d20095e","Type":"ContainerStarted","Data":"f01a4ff744e79040f9c3eec5572979e88fe32df7c77385c94a9d6c9ca006f214"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.785357 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" event={"ID":"468b20da-e9e1-43e5-a265-54d131b1b66a","Type":"ContainerStarted","Data":"2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.785482 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.798431 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.804010 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.817300 4573 generic.go:334] "Generic (PLEG): container finished" podID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerID="84010d0a1dd1d90a422b835b40ba3ca6ed6719c991a72e1cfdb2b5093ea70b08" exitCode=0 Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.817350 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bdk6" event={"ID":"bdf17da2-3e3a-4940-ae7c-680e6af80d09","Type":"ContainerDied","Data":"84010d0a1dd1d90a422b835b40ba3ca6ed6719c991a72e1cfdb2b5093ea70b08"} Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.891402 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.891387306 podStartE2EDuration="3.891387306s" podCreationTimestamp="2025-12-03 08:40:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:36.889789744 +0000 UTC m=+157.458169003" watchObservedRunningTime="2025-12-03 08:40:36.891387306 +0000 UTC m=+157.459766565" Dec 03 08:40:36 crc kubenswrapper[4573]: I1203 08:40:36.963689 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" podStartSLOduration=136.963670829 podStartE2EDuration="2m16.963670829s" podCreationTimestamp="2025-12-03 08:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:36.962631771 +0000 UTC m=+157.531011050" watchObservedRunningTime="2025-12-03 08:40:36.963670829 +0000 UTC m=+157.532050088" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.286944 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gxr7l"] Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.287537 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:37 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:37 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:37 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.287579 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.293342 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.296407 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.309479 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gxr7l"] Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.576324 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljrp9\" (UniqueName: \"kubernetes.io/projected/8920781e-296c-4525-8074-99b4d9262e6a-kube-api-access-ljrp9\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.576379 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-utilities\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.576434 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-catalog-content\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.679672 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-utilities\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.679770 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-catalog-content\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.679824 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljrp9\" (UniqueName: \"kubernetes.io/projected/8920781e-296c-4525-8074-99b4d9262e6a-kube-api-access-ljrp9\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.680864 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-utilities\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.681260 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-catalog-content\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.693867 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fzg9q"] Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.708750 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-htmvp"] Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.711332 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.752959 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-htmvp"] Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.768118 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljrp9\" (UniqueName: \"kubernetes.io/projected/8920781e-296c-4525-8074-99b4d9262e6a-kube-api-access-ljrp9\") pod \"redhat-operators-gxr7l\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.808393 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.818479 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.876181 4573 generic.go:334] "Generic (PLEG): container finished" podID="d2d9abac-8816-4343-979e-805a0d20095e" containerID="f01a4ff744e79040f9c3eec5572979e88fe32df7c77385c94a9d6c9ca006f214" exitCode=0 Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.876235 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d2d9abac-8816-4343-979e-805a0d20095e","Type":"ContainerDied","Data":"f01a4ff744e79040f9c3eec5572979e88fe32df7c77385c94a9d6c9ca006f214"} Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.876297 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kndk9"] Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.884707 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-catalog-content\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.884806 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-utilities\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.884898 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6n8c\" (UniqueName: \"kubernetes.io/projected/cddd1b3a-1843-49e2-a538-051f6f698cd4-kube-api-access-q6n8c\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.986774 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6n8c\" (UniqueName: \"kubernetes.io/projected/cddd1b3a-1843-49e2-a538-051f6f698cd4-kube-api-access-q6n8c\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.986831 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-catalog-content\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.986870 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-utilities\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.989275 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-catalog-content\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:37 crc kubenswrapper[4573]: I1203 08:40:37.989905 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-utilities\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.015614 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6n8c\" (UniqueName: \"kubernetes.io/projected/cddd1b3a-1843-49e2-a538-051f6f698cd4-kube-api-access-q6n8c\") pod \"redhat-operators-htmvp\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.121696 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.292447 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:38 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:38 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:38 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.292500 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.567931 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.578423 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gxr7l"] Dec 03 08:40:38 crc kubenswrapper[4573]: W1203 08:40:38.621190 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8920781e_296c_4525_8074_99b4d9262e6a.slice/crio-7a497d1c01f753b02f8c0c1850fed6f0c659b05b0efc62f884947c4a49ff8c05 WatchSource:0}: Error finding container 7a497d1c01f753b02f8c0c1850fed6f0c659b05b0efc62f884947c4a49ff8c05: Status 404 returned error can't find the container with id 7a497d1c01f753b02f8c0c1850fed6f0c659b05b0efc62f884947c4a49ff8c05 Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.706651 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htg6b\" (UniqueName: \"kubernetes.io/projected/2c10cae9-fe93-4134-ae2f-7e242201dc5f-kube-api-access-htg6b\") pod \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.706733 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c10cae9-fe93-4134-ae2f-7e242201dc5f-config-volume\") pod \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.706785 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c10cae9-fe93-4134-ae2f-7e242201dc5f-secret-volume\") pod \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\" (UID: \"2c10cae9-fe93-4134-ae2f-7e242201dc5f\") " Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.708769 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c10cae9-fe93-4134-ae2f-7e242201dc5f-config-volume" (OuterVolumeSpecName: "config-volume") pod "2c10cae9-fe93-4134-ae2f-7e242201dc5f" (UID: "2c10cae9-fe93-4134-ae2f-7e242201dc5f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.728794 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c10cae9-fe93-4134-ae2f-7e242201dc5f-kube-api-access-htg6b" (OuterVolumeSpecName: "kube-api-access-htg6b") pod "2c10cae9-fe93-4134-ae2f-7e242201dc5f" (UID: "2c10cae9-fe93-4134-ae2f-7e242201dc5f"). InnerVolumeSpecName "kube-api-access-htg6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.728839 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c10cae9-fe93-4134-ae2f-7e242201dc5f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2c10cae9-fe93-4134-ae2f-7e242201dc5f" (UID: "2c10cae9-fe93-4134-ae2f-7e242201dc5f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.808834 4573 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c10cae9-fe93-4134-ae2f-7e242201dc5f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.808872 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htg6b\" (UniqueName: \"kubernetes.io/projected/2c10cae9-fe93-4134-ae2f-7e242201dc5f-kube-api-access-htg6b\") on node \"crc\" DevicePath \"\"" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.808882 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c10cae9-fe93-4134-ae2f-7e242201dc5f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.982454 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxr7l" event={"ID":"8920781e-296c-4525-8074-99b4d9262e6a","Type":"ContainerStarted","Data":"7a497d1c01f753b02f8c0c1850fed6f0c659b05b0efc62f884947c4a49ff8c05"} Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.986172 4573 generic.go:334] "Generic (PLEG): container finished" podID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerID="d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153" exitCode=0 Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.986262 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fzg9q" event={"ID":"91e26815-0b3b-4855-be0a-603e52a2d02e","Type":"ContainerDied","Data":"d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153"} Dec 03 08:40:38 crc kubenswrapper[4573]: I1203 08:40:38.986309 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fzg9q" event={"ID":"91e26815-0b3b-4855-be0a-603e52a2d02e","Type":"ContainerStarted","Data":"2b9b4ed36bf8faf4a074ddf0b1d67fd4e193381183430c9954c7886de27feb52"} Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.004522 4573 generic.go:334] "Generic (PLEG): container finished" podID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerID="1f609f98771d8cedaadc125e0f4074666602e283e5d4efb87fd0cf9d5308b4d7" exitCode=0 Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.004646 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kndk9" event={"ID":"5b352afe-f1f8-4b25-b6a0-3ec01a374536","Type":"ContainerDied","Data":"1f609f98771d8cedaadc125e0f4074666602e283e5d4efb87fd0cf9d5308b4d7"} Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.004689 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kndk9" event={"ID":"5b352afe-f1f8-4b25-b6a0-3ec01a374536","Type":"ContainerStarted","Data":"746849144c4cf7b5b1e2cc27f33f821bac37d7e425d972607e09258b9430b8f3"} Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.010127 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" event={"ID":"2c10cae9-fe93-4134-ae2f-7e242201dc5f","Type":"ContainerDied","Data":"8629d95bf35528b1b982bbca877333d0cf69debb557ebcd562a2378a3f5c4991"} Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.010161 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8629d95bf35528b1b982bbca877333d0cf69debb557ebcd562a2378a3f5c4991" Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.010217 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs" Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.047939 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c","Type":"ContainerStarted","Data":"2a45000648f069126334bbfe388abde34d8d62bc736193fe76aefedca882b440"} Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.173975 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-htmvp"] Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.205063 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:39 crc kubenswrapper[4573]: W1203 08:40:39.243078 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcddd1b3a_1843_49e2_a538_051f6f698cd4.slice/crio-2bd9e52e21536b97cfe3c8c68dc5a8451dc692be86212c854461f8058d79e5ee WatchSource:0}: Error finding container 2bd9e52e21536b97cfe3c8c68dc5a8451dc692be86212c854461f8058d79e5ee: Status 404 returned error can't find the container with id 2bd9e52e21536b97cfe3c8c68dc5a8451dc692be86212c854461f8058d79e5ee Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.266416 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-fg658" Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.285369 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:39 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:39 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:39 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.285451 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.348210 4573 patch_prober.go:28] interesting pod/console-f9d7485db-dpwlm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.348262 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-dpwlm" podUID="bdc5518a-70d7-4f7f-bb7a-3320cb61f607" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 03 08:40:39 crc kubenswrapper[4573]: I1203 08:40:39.436856 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-92rk6" Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.093430 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.142796 4573 generic.go:334] "Generic (PLEG): container finished" podID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerID="758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081" exitCode=0 Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.142986 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htmvp" event={"ID":"cddd1b3a-1843-49e2-a538-051f6f698cd4","Type":"ContainerDied","Data":"758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081"} Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.143021 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htmvp" event={"ID":"cddd1b3a-1843-49e2-a538-051f6f698cd4","Type":"ContainerStarted","Data":"2bd9e52e21536b97cfe3c8c68dc5a8451dc692be86212c854461f8058d79e5ee"} Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.159108 4573 generic.go:334] "Generic (PLEG): container finished" podID="e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c" containerID="dd939c91ef4784a9dd421c23dccfdfc18661e92a104536150f802132b47d59e4" exitCode=0 Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.159504 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c","Type":"ContainerDied","Data":"dd939c91ef4784a9dd421c23dccfdfc18661e92a104536150f802132b47d59e4"} Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.164298 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2d9abac-8816-4343-979e-805a0d20095e-kubelet-dir\") pod \"d2d9abac-8816-4343-979e-805a0d20095e\" (UID: \"d2d9abac-8816-4343-979e-805a0d20095e\") " Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.170136 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2d9abac-8816-4343-979e-805a0d20095e-kube-api-access\") pod \"d2d9abac-8816-4343-979e-805a0d20095e\" (UID: \"d2d9abac-8816-4343-979e-805a0d20095e\") " Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.164457 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2d9abac-8816-4343-979e-805a0d20095e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d2d9abac-8816-4343-979e-805a0d20095e" (UID: "d2d9abac-8816-4343-979e-805a0d20095e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.170775 4573 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d2d9abac-8816-4343-979e-805a0d20095e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.228993 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d9abac-8816-4343-979e-805a0d20095e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d2d9abac-8816-4343-979e-805a0d20095e" (UID: "d2d9abac-8816-4343-979e-805a0d20095e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.229198 4573 generic.go:334] "Generic (PLEG): container finished" podID="8920781e-296c-4525-8074-99b4d9262e6a" containerID="8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8" exitCode=0 Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.229327 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxr7l" event={"ID":"8920781e-296c-4525-8074-99b4d9262e6a","Type":"ContainerDied","Data":"8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8"} Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.244413 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d2d9abac-8816-4343-979e-805a0d20095e","Type":"ContainerDied","Data":"b54c8625aefaedfde1d63f2882c2ac67434af3e70686954141f94584fc4fa558"} Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.244459 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b54c8625aefaedfde1d63f2882c2ac67434af3e70686954141f94584fc4fa558" Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.244526 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.271516 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2d9abac-8816-4343-979e-805a0d20095e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.283616 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:40 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:40 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:40 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:40 crc kubenswrapper[4573]: I1203 08:40:40.283662 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.161329 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.161403 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.166538 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.166717 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.281479 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:41 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:41 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:41 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.281730 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.334455 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.635929 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.678324 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4rx5b" Dec 03 08:40:41 crc kubenswrapper[4573]: I1203 08:40:41.901396 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.040582 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kubelet-dir\") pod \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\" (UID: \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\") " Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.040651 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kube-api-access\") pod \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\" (UID: \"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c\") " Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.041637 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c" (UID: "e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.150641 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c" (UID: "e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.151442 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.151472 4573 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.286678 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 08:40:42 crc kubenswrapper[4573]: [-]has-synced failed: reason withheld Dec 03 08:40:42 crc kubenswrapper[4573]: [+]process-running ok Dec 03 08:40:42 crc kubenswrapper[4573]: healthz check failed Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.286733 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.298380 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c","Type":"ContainerDied","Data":"2a45000648f069126334bbfe388abde34d8d62bc736193fe76aefedca882b440"} Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.298736 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a45000648f069126334bbfe388abde34d8d62bc736193fe76aefedca882b440" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.298812 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.636699 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.663002 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bb7f6603-0f41-4e0f-b77e-47dbcc676aaf-metrics-certs\") pod \"network-metrics-daemon-kcphv\" (UID: \"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf\") " pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:42 crc kubenswrapper[4573]: I1203 08:40:42.904376 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kcphv" Dec 03 08:40:43 crc kubenswrapper[4573]: I1203 08:40:43.282803 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:43 crc kubenswrapper[4573]: I1203 08:40:43.286864 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-8tpmr" Dec 03 08:40:43 crc kubenswrapper[4573]: I1203 08:40:43.771577 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kcphv"] Dec 03 08:40:44 crc kubenswrapper[4573]: I1203 08:40:44.394396 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kcphv" event={"ID":"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf","Type":"ContainerStarted","Data":"4a00416e53f42a750e4bf4db6d09ade81870b847df2447590b73addad0347cda"} Dec 03 08:40:45 crc kubenswrapper[4573]: I1203 08:40:45.416512 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kcphv" event={"ID":"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf","Type":"ContainerStarted","Data":"84556eee520676995e5f0ee139946fc66047c692bbaa4f3249394e03218258b8"} Dec 03 08:40:48 crc kubenswrapper[4573]: I1203 08:40:48.500842 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kcphv" event={"ID":"bb7f6603-0f41-4e0f-b77e-47dbcc676aaf","Type":"ContainerStarted","Data":"4dbb2c43df91f4fabd93c14d6c778240e4860b6d339a99326f512333e32b1dda"} Dec 03 08:40:48 crc kubenswrapper[4573]: I1203 08:40:48.525847 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kcphv" podStartSLOduration=149.525826703 podStartE2EDuration="2m29.525826703s" podCreationTimestamp="2025-12-03 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:40:48.521978222 +0000 UTC m=+169.090357501" watchObservedRunningTime="2025-12-03 08:40:48.525826703 +0000 UTC m=+169.094205962" Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.165602 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.165920 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.165964 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.166549 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"fa266013074576d456ab9ddd7e4a669addbfa76e50001ca84c176a67f601102f"} pod="openshift-console/downloads-7954f5f757-b92rg" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.166653 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" containerID="cri-o://fa266013074576d456ab9ddd7e4a669addbfa76e50001ca84c176a67f601102f" gracePeriod=2 Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.165816 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.166798 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.167132 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.167186 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.656172 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:51 crc kubenswrapper[4573]: I1203 08:40:51.659861 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:40:51 crc kubenswrapper[4573]: E1203 08:40:51.803736 4573 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3c8277f_eef7_4121_8b4e_ed0df8deda4b.slice/crio-conmon-fa266013074576d456ab9ddd7e4a669addbfa76e50001ca84c176a67f601102f.scope\": RecentStats: unable to find data in memory cache]" Dec 03 08:40:52 crc kubenswrapper[4573]: I1203 08:40:52.586942 4573 generic.go:334] "Generic (PLEG): container finished" podID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerID="fa266013074576d456ab9ddd7e4a669addbfa76e50001ca84c176a67f601102f" exitCode=0 Dec 03 08:40:52 crc kubenswrapper[4573]: I1203 08:40:52.586981 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b92rg" event={"ID":"b3c8277f-eef7-4121-8b4e-ed0df8deda4b","Type":"ContainerDied","Data":"fa266013074576d456ab9ddd7e4a669addbfa76e50001ca84c176a67f601102f"} Dec 03 08:40:54 crc kubenswrapper[4573]: I1203 08:40:54.118303 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:40:56 crc kubenswrapper[4573]: I1203 08:40:56.943422 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:40:56 crc kubenswrapper[4573]: I1203 08:40:56.943749 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:41:01 crc kubenswrapper[4573]: I1203 08:41:01.160595 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:41:01 crc kubenswrapper[4573]: I1203 08:41:01.161140 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:41:01 crc kubenswrapper[4573]: I1203 08:41:01.354272 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4jxgf" Dec 03 08:41:07 crc kubenswrapper[4573]: I1203 08:41:07.463581 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 08:41:11 crc kubenswrapper[4573]: I1203 08:41:11.161160 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:41:11 crc kubenswrapper[4573]: I1203 08:41:11.161973 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.969851 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 08:41:12 crc kubenswrapper[4573]: E1203 08:41:12.971734 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d9abac-8816-4343-979e-805a0d20095e" containerName="pruner" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.971851 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d9abac-8816-4343-979e-805a0d20095e" containerName="pruner" Dec 03 08:41:12 crc kubenswrapper[4573]: E1203 08:41:12.971934 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c10cae9-fe93-4134-ae2f-7e242201dc5f" containerName="collect-profiles" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.972000 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c10cae9-fe93-4134-ae2f-7e242201dc5f" containerName="collect-profiles" Dec 03 08:41:12 crc kubenswrapper[4573]: E1203 08:41:12.972091 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c" containerName="pruner" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.972179 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c" containerName="pruner" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.972364 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c10cae9-fe93-4134-ae2f-7e242201dc5f" containerName="collect-profiles" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.972441 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d9abac-8816-4343-979e-805a0d20095e" containerName="pruner" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.972567 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1814b6f-e38b-4b94-b1ea-e0ab3fbce61c" containerName="pruner" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.973259 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.977059 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.977083 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.987871 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:12 crc kubenswrapper[4573]: I1203 08:41:12.987933 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:13 crc kubenswrapper[4573]: I1203 08:41:13.020551 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 08:41:13 crc kubenswrapper[4573]: I1203 08:41:13.089251 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:13 crc kubenswrapper[4573]: I1203 08:41:13.089323 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:13 crc kubenswrapper[4573]: I1203 08:41:13.089407 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:13 crc kubenswrapper[4573]: I1203 08:41:13.109603 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:13 crc kubenswrapper[4573]: I1203 08:41:13.301682 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.008691 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.013433 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.018635 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.096493 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.096696 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kube-api-access\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.096740 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-var-lock\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.198430 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kube-api-access\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.201635 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-var-lock\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.201755 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.201842 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-var-lock\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.202010 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.225080 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kube-api-access\") pod \"installer-9-crc\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:17 crc kubenswrapper[4573]: I1203 08:41:17.387385 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:41:21 crc kubenswrapper[4573]: I1203 08:41:21.161286 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:41:21 crc kubenswrapper[4573]: I1203 08:41:21.161679 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:41:24 crc kubenswrapper[4573]: E1203 08:41:24.280700 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 08:41:24 crc kubenswrapper[4573]: E1203 08:41:24.281437 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-95rk5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-722wl_openshift-marketplace(e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 08:41:24 crc kubenswrapper[4573]: E1203 08:41:24.282792 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-722wl" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" Dec 03 08:41:24 crc kubenswrapper[4573]: E1203 08:41:24.295337 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 08:41:24 crc kubenswrapper[4573]: E1203 08:41:24.295462 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nbwkb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-f77fx_openshift-marketplace(e00387e6-6ebe-4c0e-8cf3-7c48059a79e9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 08:41:24 crc kubenswrapper[4573]: E1203 08:41:24.296748 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-f77fx" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" Dec 03 08:41:25 crc kubenswrapper[4573]: E1203 08:41:25.675348 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-f77fx" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" Dec 03 08:41:25 crc kubenswrapper[4573]: E1203 08:41:25.675505 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-722wl" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" Dec 03 08:41:25 crc kubenswrapper[4573]: E1203 08:41:25.741373 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 08:41:25 crc kubenswrapper[4573]: E1203 08:41:25.741824 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t2g6r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fzg9q_openshift-marketplace(91e26815-0b3b-4855-be0a-603e52a2d02e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 08:41:25 crc kubenswrapper[4573]: E1203 08:41:25.743408 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fzg9q" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" Dec 03 08:41:26 crc kubenswrapper[4573]: I1203 08:41:26.943265 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:41:26 crc kubenswrapper[4573]: I1203 08:41:26.944101 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:41:26 crc kubenswrapper[4573]: I1203 08:41:26.944207 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:41:26 crc kubenswrapper[4573]: I1203 08:41:26.945251 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:41:26 crc kubenswrapper[4573]: I1203 08:41:26.945319 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a" gracePeriod=600 Dec 03 08:41:28 crc kubenswrapper[4573]: I1203 08:41:28.005334 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a" exitCode=0 Dec 03 08:41:28 crc kubenswrapper[4573]: I1203 08:41:28.005423 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a"} Dec 03 08:41:29 crc kubenswrapper[4573]: I1203 08:41:29.934400 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m7kqr"] Dec 03 08:41:31 crc kubenswrapper[4573]: I1203 08:41:31.161452 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:41:31 crc kubenswrapper[4573]: I1203 08:41:31.161543 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:41:31 crc kubenswrapper[4573]: E1203 08:41:31.474874 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fzg9q" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" Dec 03 08:41:31 crc kubenswrapper[4573]: E1203 08:41:31.559259 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 08:41:31 crc kubenswrapper[4573]: E1203 08:41:31.560027 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ljrp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gxr7l_openshift-marketplace(8920781e-296c-4525-8074-99b4d9262e6a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 08:41:31 crc kubenswrapper[4573]: E1203 08:41:31.561419 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gxr7l" podUID="8920781e-296c-4525-8074-99b4d9262e6a" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.360597 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gxr7l" podUID="8920781e-296c-4525-8074-99b4d9262e6a" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.478418 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.478939 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnqgr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5bdk6_openshift-marketplace(bdf17da2-3e3a-4940-ae7c-680e6af80d09): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.480514 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5bdk6" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.529089 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.529264 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wmjd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-mbtqh_openshift-marketplace(93486cc3-d649-430e-8297-723a34105ef9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.530488 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-mbtqh" podUID="93486cc3-d649-430e-8297-723a34105ef9" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.548039 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.548249 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q6n8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-htmvp_openshift-marketplace(cddd1b3a-1843-49e2-a538-051f6f698cd4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.550258 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-htmvp" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.606320 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.606561 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qdfwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kndk9_openshift-marketplace(5b352afe-f1f8-4b25-b6a0-3ec01a374536): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 08:41:34 crc kubenswrapper[4573]: E1203 08:41:34.607900 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-kndk9" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" Dec 03 08:41:34 crc kubenswrapper[4573]: I1203 08:41:34.800595 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 08:41:35 crc kubenswrapper[4573]: I1203 08:41:35.066096 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a39c171a-2a27-4143-b1ac-ccdcba20fbc9","Type":"ContainerStarted","Data":"7104026b6eac6dd68fc704548dcffaaf3d513fdee78caba2400cb523e95f6fca"} Dec 03 08:41:35 crc kubenswrapper[4573]: I1203 08:41:35.069901 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b92rg" event={"ID":"b3c8277f-eef7-4121-8b4e-ed0df8deda4b","Type":"ContainerStarted","Data":"831e7a8536e61fae1ecfe00418f5b883b5fe8cf893085dddba63cc6287291b7e"} Dec 03 08:41:35 crc kubenswrapper[4573]: I1203 08:41:35.070268 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:41:35 crc kubenswrapper[4573]: I1203 08:41:35.070637 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:41:35 crc kubenswrapper[4573]: I1203 08:41:35.070711 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:41:35 crc kubenswrapper[4573]: I1203 08:41:35.079760 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"c69f0ae1788b1621dbf7e5b9317fb691e15bc289be6362f9912d1759f42dde21"} Dec 03 08:41:35 crc kubenswrapper[4573]: E1203 08:41:35.082216 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5bdk6" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" Dec 03 08:41:35 crc kubenswrapper[4573]: E1203 08:41:35.082522 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kndk9" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" Dec 03 08:41:35 crc kubenswrapper[4573]: E1203 08:41:35.082586 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-htmvp" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" Dec 03 08:41:35 crc kubenswrapper[4573]: E1203 08:41:35.082816 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-mbtqh" podUID="93486cc3-d649-430e-8297-723a34105ef9" Dec 03 08:41:35 crc kubenswrapper[4573]: I1203 08:41:35.157100 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 08:41:35 crc kubenswrapper[4573]: W1203 08:41:35.161833 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podaac4f7f2_0e49_4dfb_b5f2_796bc907d74b.slice/crio-b7e31e7afcc2ec748810136b409fa902d6aed9a3ca76a33c06b69d518b92575d WatchSource:0}: Error finding container b7e31e7afcc2ec748810136b409fa902d6aed9a3ca76a33c06b69d518b92575d: Status 404 returned error can't find the container with id b7e31e7afcc2ec748810136b409fa902d6aed9a3ca76a33c06b69d518b92575d Dec 03 08:41:36 crc kubenswrapper[4573]: I1203 08:41:36.086704 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b","Type":"ContainerStarted","Data":"b7e31e7afcc2ec748810136b409fa902d6aed9a3ca76a33c06b69d518b92575d"} Dec 03 08:41:36 crc kubenswrapper[4573]: I1203 08:41:36.088857 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:41:36 crc kubenswrapper[4573]: I1203 08:41:36.089013 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:41:37 crc kubenswrapper[4573]: I1203 08:41:37.096082 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a39c171a-2a27-4143-b1ac-ccdcba20fbc9","Type":"ContainerStarted","Data":"bd71edfe2851f9099b5949464c93fc83c8cf587cffeed93a9c939e23bd22589f"} Dec 03 08:41:37 crc kubenswrapper[4573]: I1203 08:41:37.100654 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b","Type":"ContainerStarted","Data":"42613390ed422c6b1b138dab2b130685b5f51d40f149042268c6179cefad83f6"} Dec 03 08:41:37 crc kubenswrapper[4573]: I1203 08:41:37.122266 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=21.122248176 podStartE2EDuration="21.122248176s" podCreationTimestamp="2025-12-03 08:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:41:37.120599372 +0000 UTC m=+217.688978662" watchObservedRunningTime="2025-12-03 08:41:37.122248176 +0000 UTC m=+217.690627425" Dec 03 08:41:37 crc kubenswrapper[4573]: I1203 08:41:37.144027 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=25.143993582 podStartE2EDuration="25.143993582s" podCreationTimestamp="2025-12-03 08:41:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:41:37.141381842 +0000 UTC m=+217.709761101" watchObservedRunningTime="2025-12-03 08:41:37.143993582 +0000 UTC m=+217.712372841" Dec 03 08:41:38 crc kubenswrapper[4573]: I1203 08:41:38.109756 4573 generic.go:334] "Generic (PLEG): container finished" podID="aac4f7f2-0e49-4dfb-b5f2-796bc907d74b" containerID="42613390ed422c6b1b138dab2b130685b5f51d40f149042268c6179cefad83f6" exitCode=0 Dec 03 08:41:38 crc kubenswrapper[4573]: I1203 08:41:38.109829 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b","Type":"ContainerDied","Data":"42613390ed422c6b1b138dab2b130685b5f51d40f149042268c6179cefad83f6"} Dec 03 08:41:39 crc kubenswrapper[4573]: I1203 08:41:39.701190 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:39 crc kubenswrapper[4573]: I1203 08:41:39.775541 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kubelet-dir\") pod \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\" (UID: \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\") " Dec 03 08:41:39 crc kubenswrapper[4573]: I1203 08:41:39.775619 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kube-api-access\") pod \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\" (UID: \"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b\") " Dec 03 08:41:39 crc kubenswrapper[4573]: I1203 08:41:39.775908 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "aac4f7f2-0e49-4dfb-b5f2-796bc907d74b" (UID: "aac4f7f2-0e49-4dfb-b5f2-796bc907d74b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:41:39 crc kubenswrapper[4573]: I1203 08:41:39.783089 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "aac4f7f2-0e49-4dfb-b5f2-796bc907d74b" (UID: "aac4f7f2-0e49-4dfb-b5f2-796bc907d74b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:41:39 crc kubenswrapper[4573]: I1203 08:41:39.876965 4573 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:39 crc kubenswrapper[4573]: I1203 08:41:39.877256 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aac4f7f2-0e49-4dfb-b5f2-796bc907d74b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:40 crc kubenswrapper[4573]: I1203 08:41:40.121561 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"aac4f7f2-0e49-4dfb-b5f2-796bc907d74b","Type":"ContainerDied","Data":"b7e31e7afcc2ec748810136b409fa902d6aed9a3ca76a33c06b69d518b92575d"} Dec 03 08:41:40 crc kubenswrapper[4573]: I1203 08:41:40.121617 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7e31e7afcc2ec748810136b409fa902d6aed9a3ca76a33c06b69d518b92575d" Dec 03 08:41:40 crc kubenswrapper[4573]: I1203 08:41:40.121685 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 08:41:41 crc kubenswrapper[4573]: I1203 08:41:41.160839 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:41:41 crc kubenswrapper[4573]: I1203 08:41:41.161134 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:41:41 crc kubenswrapper[4573]: I1203 08:41:41.163226 4573 patch_prober.go:28] interesting pod/downloads-7954f5f757-b92rg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Dec 03 08:41:41 crc kubenswrapper[4573]: I1203 08:41:41.163321 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b92rg" podUID="b3c8277f-eef7-4121-8b4e-ed0df8deda4b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Dec 03 08:41:45 crc kubenswrapper[4573]: I1203 08:41:45.156157 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-722wl" event={"ID":"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf","Type":"ContainerStarted","Data":"c3ab589892625d9410c779c47e4677108500f10d7d9213ff18e6869bab567d00"} Dec 03 08:41:46 crc kubenswrapper[4573]: I1203 08:41:46.166412 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f77fx" event={"ID":"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9","Type":"ContainerStarted","Data":"4ce9b1d469f174a004233e1f2779bf1f5071dae8d853ba689728f9e428d45fb9"} Dec 03 08:41:47 crc kubenswrapper[4573]: I1203 08:41:47.199241 4573 generic.go:334] "Generic (PLEG): container finished" podID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerID="c3ab589892625d9410c779c47e4677108500f10d7d9213ff18e6869bab567d00" exitCode=0 Dec 03 08:41:47 crc kubenswrapper[4573]: I1203 08:41:47.199960 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-722wl" event={"ID":"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf","Type":"ContainerDied","Data":"c3ab589892625d9410c779c47e4677108500f10d7d9213ff18e6869bab567d00"} Dec 03 08:41:47 crc kubenswrapper[4573]: I1203 08:41:47.208410 4573 generic.go:334] "Generic (PLEG): container finished" podID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerID="4ce9b1d469f174a004233e1f2779bf1f5071dae8d853ba689728f9e428d45fb9" exitCode=0 Dec 03 08:41:47 crc kubenswrapper[4573]: I1203 08:41:47.208500 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f77fx" event={"ID":"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9","Type":"ContainerDied","Data":"4ce9b1d469f174a004233e1f2779bf1f5071dae8d853ba689728f9e428d45fb9"} Dec 03 08:41:51 crc kubenswrapper[4573]: I1203 08:41:51.180305 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-b92rg" Dec 03 08:41:52 crc kubenswrapper[4573]: I1203 08:41:52.267619 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f77fx" event={"ID":"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9","Type":"ContainerStarted","Data":"e9b4ed2c6b262d441fddcd3e50a839dc7df58ce9f4fcb8e9fac32872dbcdc59c"} Dec 03 08:41:52 crc kubenswrapper[4573]: I1203 08:41:52.272953 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fzg9q" event={"ID":"91e26815-0b3b-4855-be0a-603e52a2d02e","Type":"ContainerStarted","Data":"24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76"} Dec 03 08:41:52 crc kubenswrapper[4573]: I1203 08:41:52.282461 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-722wl" event={"ID":"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf","Type":"ContainerStarted","Data":"750dddb0619e98e0d167fb09b34b3ce1b7fbd74f3dedcd97e57bea54d48ea67a"} Dec 03 08:41:52 crc kubenswrapper[4573]: I1203 08:41:52.286396 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kndk9" event={"ID":"5b352afe-f1f8-4b25-b6a0-3ec01a374536","Type":"ContainerStarted","Data":"47497fb77aa3eac93f428403ab554d106e254fc8ee5af3d19e7e67bba7f57fa7"} Dec 03 08:41:52 crc kubenswrapper[4573]: I1203 08:41:52.303501 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f77fx" podStartSLOduration=3.075879724 podStartE2EDuration="1m18.303471167s" podCreationTimestamp="2025-12-03 08:40:34 +0000 UTC" firstStartedPulling="2025-12-03 08:40:36.727145505 +0000 UTC m=+157.295524764" lastFinishedPulling="2025-12-03 08:41:51.954736948 +0000 UTC m=+232.523116207" observedRunningTime="2025-12-03 08:41:52.298294668 +0000 UTC m=+232.866673927" watchObservedRunningTime="2025-12-03 08:41:52.303471167 +0000 UTC m=+232.871850426" Dec 03 08:41:52 crc kubenswrapper[4573]: I1203 08:41:52.320635 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-722wl" podStartSLOduration=3.180356121 podStartE2EDuration="1m18.320608409s" podCreationTimestamp="2025-12-03 08:40:34 +0000 UTC" firstStartedPulling="2025-12-03 08:40:36.711317722 +0000 UTC m=+157.279696981" lastFinishedPulling="2025-12-03 08:41:51.85157001 +0000 UTC m=+232.419949269" observedRunningTime="2025-12-03 08:41:52.318018789 +0000 UTC m=+232.886398068" watchObservedRunningTime="2025-12-03 08:41:52.320608409 +0000 UTC m=+232.888987668" Dec 03 08:41:54 crc kubenswrapper[4573]: I1203 08:41:54.307176 4573 generic.go:334] "Generic (PLEG): container finished" podID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerID="47497fb77aa3eac93f428403ab554d106e254fc8ee5af3d19e7e67bba7f57fa7" exitCode=0 Dec 03 08:41:54 crc kubenswrapper[4573]: I1203 08:41:54.307556 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kndk9" event={"ID":"5b352afe-f1f8-4b25-b6a0-3ec01a374536","Type":"ContainerDied","Data":"47497fb77aa3eac93f428403ab554d106e254fc8ee5af3d19e7e67bba7f57fa7"} Dec 03 08:41:54 crc kubenswrapper[4573]: I1203 08:41:54.703717 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-722wl" Dec 03 08:41:54 crc kubenswrapper[4573]: I1203 08:41:54.704076 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-722wl" Dec 03 08:41:55 crc kubenswrapper[4573]: I1203 08:41:55.001137 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" podUID="2cf8e819-4c7b-41c6-984c-871a6ac87428" containerName="oauth-openshift" containerID="cri-o://746c799e858274da8a76a3eb994ab6f01eef94ca3668977ce7485b73ec7f3f21" gracePeriod=15 Dec 03 08:41:55 crc kubenswrapper[4573]: I1203 08:41:55.055231 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:41:55 crc kubenswrapper[4573]: I1203 08:41:55.055561 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.384951 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxr7l" event={"ID":"8920781e-296c-4525-8074-99b4d9262e6a","Type":"ContainerStarted","Data":"a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d"} Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.386599 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-f77fx" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="registry-server" probeResult="failure" output=< Dec 03 08:41:56 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 08:41:56 crc kubenswrapper[4573]: > Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.393790 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-722wl" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="registry-server" probeResult="failure" output=< Dec 03 08:41:56 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 08:41:56 crc kubenswrapper[4573]: > Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.394530 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bdk6" event={"ID":"bdf17da2-3e3a-4940-ae7c-680e6af80d09","Type":"ContainerStarted","Data":"f82d0601ad80b179fe25c350291e5435102f146c54c3ce30c7c1125edfcc7d84"} Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.403183 4573 generic.go:334] "Generic (PLEG): container finished" podID="2cf8e819-4c7b-41c6-984c-871a6ac87428" containerID="746c799e858274da8a76a3eb994ab6f01eef94ca3668977ce7485b73ec7f3f21" exitCode=0 Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.403256 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" event={"ID":"2cf8e819-4c7b-41c6-984c-871a6ac87428","Type":"ContainerDied","Data":"746c799e858274da8a76a3eb994ab6f01eef94ca3668977ce7485b73ec7f3f21"} Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.406488 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbtqh" event={"ID":"93486cc3-d649-430e-8297-723a34105ef9","Type":"ContainerStarted","Data":"3c608fd699b12196fa644e4b4b23c95ea042bd27ee059a0eab50fcaf5dc44d86"} Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.408840 4573 generic.go:334] "Generic (PLEG): container finished" podID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerID="24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76" exitCode=0 Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.408922 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fzg9q" event={"ID":"91e26815-0b3b-4855-be0a-603e52a2d02e","Type":"ContainerDied","Data":"24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76"} Dec 03 08:41:56 crc kubenswrapper[4573]: I1203 08:41:56.507407 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htmvp" event={"ID":"cddd1b3a-1843-49e2-a538-051f6f698cd4","Type":"ContainerStarted","Data":"83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969"} Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.075379 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.151943 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg"] Dec 03 08:41:57 crc kubenswrapper[4573]: E1203 08:41:57.152217 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aac4f7f2-0e49-4dfb-b5f2-796bc907d74b" containerName="pruner" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.152236 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac4f7f2-0e49-4dfb-b5f2-796bc907d74b" containerName="pruner" Dec 03 08:41:57 crc kubenswrapper[4573]: E1203 08:41:57.152254 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf8e819-4c7b-41c6-984c-871a6ac87428" containerName="oauth-openshift" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.152263 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf8e819-4c7b-41c6-984c-871a6ac87428" containerName="oauth-openshift" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.152396 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf8e819-4c7b-41c6-984c-871a6ac87428" containerName="oauth-openshift" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.152419 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="aac4f7f2-0e49-4dfb-b5f2-796bc907d74b" containerName="pruner" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.152863 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207638 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-error\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207689 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-trusted-ca-bundle\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207734 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-session\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207779 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-router-certs\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207885 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-serving-cert\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207918 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-login\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207938 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-provider-selection\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207973 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-ocp-branding-template\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.207997 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-idp-0-file-data\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208017 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-dir\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208080 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-cliconfig\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208116 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf78f\" (UniqueName: \"kubernetes.io/projected/2cf8e819-4c7b-41c6-984c-871a6ac87428-kube-api-access-gf78f\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208140 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-service-ca\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208163 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-policies\") pod \"2cf8e819-4c7b-41c6-984c-871a6ac87428\" (UID: \"2cf8e819-4c7b-41c6-984c-871a6ac87428\") " Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208388 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208439 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208467 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-audit-policies\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208490 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208519 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-error\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208539 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pknxn\" (UniqueName: \"kubernetes.io/projected/7da099a6-df67-490b-b5ab-2cb6cc44e438-kube-api-access-pknxn\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208561 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208584 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-login\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208605 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-service-ca\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208632 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208658 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208697 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7da099a6-df67-490b-b5ab-2cb6cc44e438-audit-dir\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208721 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-session\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208753 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-router-certs\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.208909 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.211847 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.211920 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.213829 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.228542 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.232488 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cf8e819-4c7b-41c6-984c-871a6ac87428-kube-api-access-gf78f" (OuterVolumeSpecName: "kube-api-access-gf78f") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "kube-api-access-gf78f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.236407 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.241461 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.242878 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.267734 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.278519 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.285888 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.286550 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.287662 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg"] Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.296620 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2cf8e819-4c7b-41c6-984c-871a6ac87428" (UID: "2cf8e819-4c7b-41c6-984c-871a6ac87428"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321023 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321230 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321311 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7da099a6-df67-490b-b5ab-2cb6cc44e438-audit-dir\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321444 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-session\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321536 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-router-certs\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321591 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321680 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321726 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-audit-policies\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321754 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321794 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-error\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321821 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pknxn\" (UniqueName: \"kubernetes.io/projected/7da099a6-df67-490b-b5ab-2cb6cc44e438-kube-api-access-pknxn\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321857 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321889 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-login\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321915 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-service-ca\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.321998 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322013 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322027 4573 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322038 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322134 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf78f\" (UniqueName: \"kubernetes.io/projected/2cf8e819-4c7b-41c6-984c-871a6ac87428-kube-api-access-gf78f\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322149 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322161 4573 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322191 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322202 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322215 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322225 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322234 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322244 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322256 4573 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2cf8e819-4c7b-41c6-984c-871a6ac87428-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.322504 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7da099a6-df67-490b-b5ab-2cb6cc44e438-audit-dir\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.323221 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-service-ca\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.323795 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.324305 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-audit-policies\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.325893 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.331474 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-session\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.335242 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-error\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.341535 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.342011 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-router-certs\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.343648 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-login\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.346882 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.357713 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.358234 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7da099a6-df67-490b-b5ab-2cb6cc44e438-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.396549 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pknxn\" (UniqueName: \"kubernetes.io/projected/7da099a6-df67-490b-b5ab-2cb6cc44e438-kube-api-access-pknxn\") pod \"oauth-openshift-7b4b9565b9-sjwxg\" (UID: \"7da099a6-df67-490b-b5ab-2cb6cc44e438\") " pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.502721 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.521759 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.529423 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-m7kqr" event={"ID":"2cf8e819-4c7b-41c6-984c-871a6ac87428","Type":"ContainerDied","Data":"a27a2d3b3d44c72995cfce373fcec5d69655d2d4ac41c2c3c698ab65ffda8f57"} Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.529536 4573 scope.go:117] "RemoveContainer" containerID="746c799e858274da8a76a3eb994ab6f01eef94ca3668977ce7485b73ec7f3f21" Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.538026 4573 generic.go:334] "Generic (PLEG): container finished" podID="93486cc3-d649-430e-8297-723a34105ef9" containerID="3c608fd699b12196fa644e4b4b23c95ea042bd27ee059a0eab50fcaf5dc44d86" exitCode=0 Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.538185 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbtqh" event={"ID":"93486cc3-d649-430e-8297-723a34105ef9","Type":"ContainerDied","Data":"3c608fd699b12196fa644e4b4b23c95ea042bd27ee059a0eab50fcaf5dc44d86"} Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.552154 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kndk9" event={"ID":"5b352afe-f1f8-4b25-b6a0-3ec01a374536","Type":"ContainerStarted","Data":"8be50b8e9590b42f20265a6cadaf291c65fa9a0a7ae17a263271245423b7d046"} Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.727874 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m7kqr"] Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.736611 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-m7kqr"] Dec 03 08:41:57 crc kubenswrapper[4573]: I1203 08:41:57.768363 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kndk9" podStartSLOduration=4.037149277 podStartE2EDuration="1m21.768333538s" podCreationTimestamp="2025-12-03 08:40:36 +0000 UTC" firstStartedPulling="2025-12-03 08:40:39.03001399 +0000 UTC m=+159.598393249" lastFinishedPulling="2025-12-03 08:41:56.761198251 +0000 UTC m=+237.329577510" observedRunningTime="2025-12-03 08:41:57.767589938 +0000 UTC m=+238.335969197" watchObservedRunningTime="2025-12-03 08:41:57.768333538 +0000 UTC m=+238.336712797" Dec 03 08:41:58 crc kubenswrapper[4573]: I1203 08:41:58.037907 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cf8e819-4c7b-41c6-984c-871a6ac87428" path="/var/lib/kubelet/pods/2cf8e819-4c7b-41c6-984c-871a6ac87428/volumes" Dec 03 08:41:59 crc kubenswrapper[4573]: I1203 08:41:59.017847 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fzg9q" event={"ID":"91e26815-0b3b-4855-be0a-603e52a2d02e","Type":"ContainerStarted","Data":"4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b"} Dec 03 08:41:59 crc kubenswrapper[4573]: I1203 08:41:59.019394 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg"] Dec 03 08:41:59 crc kubenswrapper[4573]: I1203 08:41:59.021249 4573 generic.go:334] "Generic (PLEG): container finished" podID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerID="f82d0601ad80b179fe25c350291e5435102f146c54c3ce30c7c1125edfcc7d84" exitCode=0 Dec 03 08:41:59 crc kubenswrapper[4573]: I1203 08:41:59.021310 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bdk6" event={"ID":"bdf17da2-3e3a-4940-ae7c-680e6af80d09","Type":"ContainerDied","Data":"f82d0601ad80b179fe25c350291e5435102f146c54c3ce30c7c1125edfcc7d84"} Dec 03 08:41:59 crc kubenswrapper[4573]: I1203 08:41:59.103297 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fzg9q" podStartSLOduration=4.919943237 podStartE2EDuration="1m23.103276761s" podCreationTimestamp="2025-12-03 08:40:36 +0000 UTC" firstStartedPulling="2025-12-03 08:40:38.988424941 +0000 UTC m=+159.556804200" lastFinishedPulling="2025-12-03 08:41:57.171758465 +0000 UTC m=+237.740137724" observedRunningTime="2025-12-03 08:41:59.099702866 +0000 UTC m=+239.668082125" watchObservedRunningTime="2025-12-03 08:41:59.103276761 +0000 UTC m=+239.671656020" Dec 03 08:42:00 crc kubenswrapper[4573]: I1203 08:42:00.044651 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" event={"ID":"7da099a6-df67-490b-b5ab-2cb6cc44e438","Type":"ContainerStarted","Data":"dbfd0fa6e38eeb479876951483f22cb5b80072e82d7978405b74ce11cd53e3a4"} Dec 03 08:42:00 crc kubenswrapper[4573]: I1203 08:42:00.046942 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbtqh" event={"ID":"93486cc3-d649-430e-8297-723a34105ef9","Type":"ContainerStarted","Data":"8cc00162df3e3e3034d2e7920f7d4c1300accdd1dba92641f57865f8b6b14cff"} Dec 03 08:42:01 crc kubenswrapper[4573]: I1203 08:42:01.055312 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" event={"ID":"7da099a6-df67-490b-b5ab-2cb6cc44e438","Type":"ContainerStarted","Data":"82b0f151883829c934ee1a18d8d9fe70858a71cdeee3dbf30376e44f55a8c9ec"} Dec 03 08:42:01 crc kubenswrapper[4573]: I1203 08:42:01.055602 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:42:01 crc kubenswrapper[4573]: I1203 08:42:01.059271 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bdk6" event={"ID":"bdf17da2-3e3a-4940-ae7c-680e6af80d09","Type":"ContainerStarted","Data":"23c53945a36baf5c3722738ca2da6593a037261aae027cd3a869f3ac2b627f87"} Dec 03 08:42:01 crc kubenswrapper[4573]: I1203 08:42:01.096839 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mbtqh" podStartSLOduration=5.133884588 podStartE2EDuration="1m27.096810126s" podCreationTimestamp="2025-12-03 08:40:34 +0000 UTC" firstStartedPulling="2025-12-03 08:40:36.711373533 +0000 UTC m=+157.279752792" lastFinishedPulling="2025-12-03 08:41:58.674299071 +0000 UTC m=+239.242678330" observedRunningTime="2025-12-03 08:42:00.117415207 +0000 UTC m=+240.685794476" watchObservedRunningTime="2025-12-03 08:42:01.096810126 +0000 UTC m=+241.665189395" Dec 03 08:42:01 crc kubenswrapper[4573]: I1203 08:42:01.107114 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" Dec 03 08:42:01 crc kubenswrapper[4573]: I1203 08:42:01.253682 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7b4b9565b9-sjwxg" podStartSLOduration=32.253664329 podStartE2EDuration="32.253664329s" podCreationTimestamp="2025-12-03 08:41:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:42:01.102577151 +0000 UTC m=+241.670956420" watchObservedRunningTime="2025-12-03 08:42:01.253664329 +0000 UTC m=+241.822043588" Dec 03 08:42:01 crc kubenswrapper[4573]: I1203 08:42:01.281011 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5bdk6" podStartSLOduration=4.442050841 podStartE2EDuration="1m27.280993925s" podCreationTimestamp="2025-12-03 08:40:34 +0000 UTC" firstStartedPulling="2025-12-03 08:40:36.858854544 +0000 UTC m=+157.427233803" lastFinishedPulling="2025-12-03 08:41:59.697797628 +0000 UTC m=+240.266176887" observedRunningTime="2025-12-03 08:42:01.254356549 +0000 UTC m=+241.822735808" watchObservedRunningTime="2025-12-03 08:42:01.280993925 +0000 UTC m=+241.849373184" Dec 03 08:42:03 crc kubenswrapper[4573]: I1203 08:42:03.084429 4573 generic.go:334] "Generic (PLEG): container finished" podID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerID="83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969" exitCode=0 Dec 03 08:42:03 crc kubenswrapper[4573]: I1203 08:42:03.084528 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htmvp" event={"ID":"cddd1b3a-1843-49e2-a538-051f6f698cd4","Type":"ContainerDied","Data":"83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969"} Dec 03 08:42:03 crc kubenswrapper[4573]: I1203 08:42:03.116758 4573 generic.go:334] "Generic (PLEG): container finished" podID="8920781e-296c-4525-8074-99b4d9262e6a" containerID="a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d" exitCode=0 Dec 03 08:42:03 crc kubenswrapper[4573]: I1203 08:42:03.116879 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxr7l" event={"ID":"8920781e-296c-4525-8074-99b4d9262e6a","Type":"ContainerDied","Data":"a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d"} Dec 03 08:42:04 crc kubenswrapper[4573]: I1203 08:42:04.467393 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:42:04 crc kubenswrapper[4573]: I1203 08:42:04.467993 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:42:04 crc kubenswrapper[4573]: I1203 08:42:04.528006 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:42:04 crc kubenswrapper[4573]: I1203 08:42:04.754512 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-722wl" Dec 03 08:42:04 crc kubenswrapper[4573]: I1203 08:42:04.800428 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-722wl" Dec 03 08:42:04 crc kubenswrapper[4573]: I1203 08:42:04.881165 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:42:04 crc kubenswrapper[4573]: I1203 08:42:04.881295 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:42:04 crc kubenswrapper[4573]: I1203 08:42:04.923086 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:42:05 crc kubenswrapper[4573]: I1203 08:42:05.096116 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:42:05 crc kubenswrapper[4573]: I1203 08:42:05.145720 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:42:05 crc kubenswrapper[4573]: I1203 08:42:05.173663 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:42:05 crc kubenswrapper[4573]: I1203 08:42:05.181855 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:42:06 crc kubenswrapper[4573]: I1203 08:42:06.747334 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:42:06 crc kubenswrapper[4573]: I1203 08:42:06.750334 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:42:06 crc kubenswrapper[4573]: I1203 08:42:06.802452 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbtqh"] Dec 03 08:42:06 crc kubenswrapper[4573]: I1203 08:42:06.804908 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:42:06 crc kubenswrapper[4573]: I1203 08:42:06.805017 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:42:06 crc kubenswrapper[4573]: I1203 08:42:06.821319 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:42:06 crc kubenswrapper[4573]: I1203 08:42:06.874128 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:42:07 crc kubenswrapper[4573]: I1203 08:42:07.142432 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mbtqh" podUID="93486cc3-d649-430e-8297-723a34105ef9" containerName="registry-server" containerID="cri-o://8cc00162df3e3e3034d2e7920f7d4c1300accdd1dba92641f57865f8b6b14cff" gracePeriod=2 Dec 03 08:42:07 crc kubenswrapper[4573]: I1203 08:42:07.188641 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:42:07 crc kubenswrapper[4573]: I1203 08:42:07.224261 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.149817 4573 generic.go:334] "Generic (PLEG): container finished" podID="93486cc3-d649-430e-8297-723a34105ef9" containerID="8cc00162df3e3e3034d2e7920f7d4c1300accdd1dba92641f57865f8b6b14cff" exitCode=0 Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.149844 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbtqh" event={"ID":"93486cc3-d649-430e-8297-723a34105ef9","Type":"ContainerDied","Data":"8cc00162df3e3e3034d2e7920f7d4c1300accdd1dba92641f57865f8b6b14cff"} Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.152635 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htmvp" event={"ID":"cddd1b3a-1843-49e2-a538-051f6f698cd4","Type":"ContainerStarted","Data":"f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e"} Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.177263 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-htmvp" podStartSLOduration=5.229012689 podStartE2EDuration="1m31.177237016s" podCreationTimestamp="2025-12-03 08:40:37 +0000 UTC" firstStartedPulling="2025-12-03 08:40:40.192785374 +0000 UTC m=+160.761164633" lastFinishedPulling="2025-12-03 08:42:06.141009701 +0000 UTC m=+246.709388960" observedRunningTime="2025-12-03 08:42:08.176147266 +0000 UTC m=+248.744526535" watchObservedRunningTime="2025-12-03 08:42:08.177237016 +0000 UTC m=+248.745616275" Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.201659 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f77fx"] Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.202104 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f77fx" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="registry-server" containerID="cri-o://e9b4ed2c6b262d441fddcd3e50a839dc7df58ce9f4fcb8e9fac32872dbcdc59c" gracePeriod=2 Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.679737 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.774848 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-utilities\") pod \"93486cc3-d649-430e-8297-723a34105ef9\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.774949 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmjd9\" (UniqueName: \"kubernetes.io/projected/93486cc3-d649-430e-8297-723a34105ef9-kube-api-access-wmjd9\") pod \"93486cc3-d649-430e-8297-723a34105ef9\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.774976 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-catalog-content\") pod \"93486cc3-d649-430e-8297-723a34105ef9\" (UID: \"93486cc3-d649-430e-8297-723a34105ef9\") " Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.775891 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-utilities" (OuterVolumeSpecName: "utilities") pod "93486cc3-d649-430e-8297-723a34105ef9" (UID: "93486cc3-d649-430e-8297-723a34105ef9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.795325 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93486cc3-d649-430e-8297-723a34105ef9-kube-api-access-wmjd9" (OuterVolumeSpecName: "kube-api-access-wmjd9") pod "93486cc3-d649-430e-8297-723a34105ef9" (UID: "93486cc3-d649-430e-8297-723a34105ef9"). InnerVolumeSpecName "kube-api-access-wmjd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.855257 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93486cc3-d649-430e-8297-723a34105ef9" (UID: "93486cc3-d649-430e-8297-723a34105ef9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.877979 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.878079 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmjd9\" (UniqueName: \"kubernetes.io/projected/93486cc3-d649-430e-8297-723a34105ef9-kube-api-access-wmjd9\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:08 crc kubenswrapper[4573]: I1203 08:42:08.878102 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93486cc3-d649-430e-8297-723a34105ef9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.163362 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbtqh" event={"ID":"93486cc3-d649-430e-8297-723a34105ef9","Type":"ContainerDied","Data":"370b20d43851103106d464c74631d8c3d3016edd72ecf2e46a3b54553503d80c"} Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.163434 4573 scope.go:117] "RemoveContainer" containerID="8cc00162df3e3e3034d2e7920f7d4c1300accdd1dba92641f57865f8b6b14cff" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.163608 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbtqh" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.173010 4573 generic.go:334] "Generic (PLEG): container finished" podID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerID="e9b4ed2c6b262d441fddcd3e50a839dc7df58ce9f4fcb8e9fac32872dbcdc59c" exitCode=0 Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.173153 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f77fx" event={"ID":"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9","Type":"ContainerDied","Data":"e9b4ed2c6b262d441fddcd3e50a839dc7df58ce9f4fcb8e9fac32872dbcdc59c"} Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.183754 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxr7l" event={"ID":"8920781e-296c-4525-8074-99b4d9262e6a","Type":"ContainerStarted","Data":"432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f"} Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.205247 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gxr7l" podStartSLOduration=4.507029179 podStartE2EDuration="1m32.205217984s" podCreationTimestamp="2025-12-03 08:40:37 +0000 UTC" firstStartedPulling="2025-12-03 08:40:40.23389334 +0000 UTC m=+160.802272599" lastFinishedPulling="2025-12-03 08:42:07.932082145 +0000 UTC m=+248.500461404" observedRunningTime="2025-12-03 08:42:09.202662805 +0000 UTC m=+249.771042064" watchObservedRunningTime="2025-12-03 08:42:09.205217984 +0000 UTC m=+249.773597243" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.226622 4573 scope.go:117] "RemoveContainer" containerID="3c608fd699b12196fa644e4b4b23c95ea042bd27ee059a0eab50fcaf5dc44d86" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.236415 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.239381 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbtqh"] Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.239442 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mbtqh"] Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.255475 4573 scope.go:117] "RemoveContainer" containerID="ccdc12851d76fd4700cc00c4270ab9dd3717161b09ad4e8c0b984ca0dd2b5409" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.389259 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbwkb\" (UniqueName: \"kubernetes.io/projected/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-kube-api-access-nbwkb\") pod \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.389392 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-catalog-content\") pod \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.389509 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-utilities\") pod \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\" (UID: \"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9\") " Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.390704 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-utilities" (OuterVolumeSpecName: "utilities") pod "e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" (UID: "e00387e6-6ebe-4c0e-8cf3-7c48059a79e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.395091 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-kube-api-access-nbwkb" (OuterVolumeSpecName: "kube-api-access-nbwkb") pod "e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" (UID: "e00387e6-6ebe-4c0e-8cf3-7c48059a79e9"). InnerVolumeSpecName "kube-api-access-nbwkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.452264 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" (UID: "e00387e6-6ebe-4c0e-8cf3-7c48059a79e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.494041 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.494106 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbwkb\" (UniqueName: \"kubernetes.io/projected/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-kube-api-access-nbwkb\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:09 crc kubenswrapper[4573]: I1203 08:42:09.494116 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.038238 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93486cc3-d649-430e-8297-723a34105ef9" path="/var/lib/kubelet/pods/93486cc3-d649-430e-8297-723a34105ef9/volumes" Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.192443 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f77fx" event={"ID":"e00387e6-6ebe-4c0e-8cf3-7c48059a79e9","Type":"ContainerDied","Data":"e99d8d8f966904eef2140be92db550a377cc5d704f0194ae0b2622f4cf3f2ba9"} Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.192520 4573 scope.go:117] "RemoveContainer" containerID="e9b4ed2c6b262d441fddcd3e50a839dc7df58ce9f4fcb8e9fac32872dbcdc59c" Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.192726 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f77fx" Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.212827 4573 scope.go:117] "RemoveContainer" containerID="4ce9b1d469f174a004233e1f2779bf1f5071dae8d853ba689728f9e428d45fb9" Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.216974 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f77fx"] Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.225721 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f77fx"] Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.236458 4573 scope.go:117] "RemoveContainer" containerID="b87ce9be9ba90ab57297935dfbc1bf5c9e17f8fae263242316520a725cf00d98" Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.609347 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kndk9"] Dec 03 08:42:10 crc kubenswrapper[4573]: I1203 08:42:10.609666 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kndk9" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerName="registry-server" containerID="cri-o://8be50b8e9590b42f20265a6cadaf291c65fa9a0a7ae17a263271245423b7d046" gracePeriod=2 Dec 03 08:42:11 crc kubenswrapper[4573]: I1203 08:42:11.203600 4573 generic.go:334] "Generic (PLEG): container finished" podID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerID="8be50b8e9590b42f20265a6cadaf291c65fa9a0a7ae17a263271245423b7d046" exitCode=0 Dec 03 08:42:11 crc kubenswrapper[4573]: I1203 08:42:11.203695 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kndk9" event={"ID":"5b352afe-f1f8-4b25-b6a0-3ec01a374536","Type":"ContainerDied","Data":"8be50b8e9590b42f20265a6cadaf291c65fa9a0a7ae17a263271245423b7d046"} Dec 03 08:42:11 crc kubenswrapper[4573]: I1203 08:42:11.926346 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.041173 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" path="/var/lib/kubelet/pods/e00387e6-6ebe-4c0e-8cf3-7c48059a79e9/volumes" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.060536 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdfwb\" (UniqueName: \"kubernetes.io/projected/5b352afe-f1f8-4b25-b6a0-3ec01a374536-kube-api-access-qdfwb\") pod \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.060658 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-utilities\") pod \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.060837 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-catalog-content\") pod \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\" (UID: \"5b352afe-f1f8-4b25-b6a0-3ec01a374536\") " Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.062951 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-utilities" (OuterVolumeSpecName: "utilities") pod "5b352afe-f1f8-4b25-b6a0-3ec01a374536" (UID: "5b352afe-f1f8-4b25-b6a0-3ec01a374536"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.070083 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b352afe-f1f8-4b25-b6a0-3ec01a374536-kube-api-access-qdfwb" (OuterVolumeSpecName: "kube-api-access-qdfwb") pod "5b352afe-f1f8-4b25-b6a0-3ec01a374536" (UID: "5b352afe-f1f8-4b25-b6a0-3ec01a374536"). InnerVolumeSpecName "kube-api-access-qdfwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.081825 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b352afe-f1f8-4b25-b6a0-3ec01a374536" (UID: "5b352afe-f1f8-4b25-b6a0-3ec01a374536"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.162462 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.162744 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdfwb\" (UniqueName: \"kubernetes.io/projected/5b352afe-f1f8-4b25-b6a0-3ec01a374536-kube-api-access-qdfwb\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.162760 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b352afe-f1f8-4b25-b6a0-3ec01a374536-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.214840 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kndk9" event={"ID":"5b352afe-f1f8-4b25-b6a0-3ec01a374536","Type":"ContainerDied","Data":"746849144c4cf7b5b1e2cc27f33f821bac37d7e425d972607e09258b9430b8f3"} Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.214934 4573 scope.go:117] "RemoveContainer" containerID="8be50b8e9590b42f20265a6cadaf291c65fa9a0a7ae17a263271245423b7d046" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.214873 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kndk9" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.236305 4573 scope.go:117] "RemoveContainer" containerID="47497fb77aa3eac93f428403ab554d106e254fc8ee5af3d19e7e67bba7f57fa7" Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.247029 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kndk9"] Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.251629 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kndk9"] Dec 03 08:42:12 crc kubenswrapper[4573]: I1203 08:42:12.269873 4573 scope.go:117] "RemoveContainer" containerID="1f609f98771d8cedaadc125e0f4074666602e283e5d4efb87fd0cf9d5308b4d7" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.037807 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" path="/var/lib/kubelet/pods/5b352afe-f1f8-4b25-b6a0-3ec01a374536/volumes" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.162865 4573 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163524 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerName="extract-utilities" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163540 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerName="extract-utilities" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163561 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93486cc3-d649-430e-8297-723a34105ef9" containerName="extract-utilities" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163569 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="93486cc3-d649-430e-8297-723a34105ef9" containerName="extract-utilities" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163578 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163587 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163596 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="extract-utilities" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163602 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="extract-utilities" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163614 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163620 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163631 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerName="extract-content" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163640 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerName="extract-content" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163650 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93486cc3-d649-430e-8297-723a34105ef9" containerName="extract-content" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163658 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="93486cc3-d649-430e-8297-723a34105ef9" containerName="extract-content" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163669 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93486cc3-d649-430e-8297-723a34105ef9" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163675 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="93486cc3-d649-430e-8297-723a34105ef9" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.163683 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="extract-content" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163692 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="extract-content" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163805 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="93486cc3-d649-430e-8297-723a34105ef9" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163816 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b352afe-f1f8-4b25-b6a0-3ec01a374536" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.163824 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="e00387e6-6ebe-4c0e-8cf3-7c48059a79e9" containerName="registry-server" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.164271 4573 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.164624 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97" gracePeriod=15 Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.164772 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4" gracePeriod=15 Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.164708 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9" gracePeriod=15 Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.164795 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.164720 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb" gracePeriod=15 Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.164854 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57" gracePeriod=15 Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.165890 4573 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.166236 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166257 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.166284 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166292 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.166305 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166312 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.166324 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166333 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.166345 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166355 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.166364 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166374 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166530 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166545 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166554 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166565 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166575 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166582 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.166719 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.166729 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.211442 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.301765 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.301846 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.301886 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.301906 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.302065 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.302099 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.302138 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.302161 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403268 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403344 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403380 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403401 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403476 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403577 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403585 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403694 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403742 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403801 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403838 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403856 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403884 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403889 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403923 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.403990 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: I1203 08:42:14.507071 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:42:14 crc kubenswrapper[4573]: W1203 08:42:14.534662 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-0e6cb2418638b487d63a749d56008cd7840dc6ee287eb1c9ea368cafc152aabe WatchSource:0}: Error finding container 0e6cb2418638b487d63a749d56008cd7840dc6ee287eb1c9ea368cafc152aabe: Status 404 returned error can't find the container with id 0e6cb2418638b487d63a749d56008cd7840dc6ee287eb1c9ea368cafc152aabe Dec 03 08:42:14 crc kubenswrapper[4573]: E1203 08:42:14.856126 4573 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187da7fd2ef38c10 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 08:42:14.855379984 +0000 UTC m=+255.423759243,LastTimestamp:2025-12-03 08:42:14.855379984 +0000 UTC m=+255.423759243,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.241514 4573 generic.go:334] "Generic (PLEG): container finished" podID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" containerID="bd71edfe2851f9099b5949464c93fc83c8cf587cffeed93a9c939e23bd22589f" exitCode=0 Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.241619 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a39c171a-2a27-4143-b1ac-ccdcba20fbc9","Type":"ContainerDied","Data":"bd71edfe2851f9099b5949464c93fc83c8cf587cffeed93a9c939e23bd22589f"} Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.244728 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.245392 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.246030 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.246982 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409"} Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.247021 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0e6cb2418638b487d63a749d56008cd7840dc6ee287eb1c9ea368cafc152aabe"} Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.247528 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.247841 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.248313 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.251205 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.252896 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.253789 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb" exitCode=0 Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.253819 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4" exitCode=0 Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.253836 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9" exitCode=0 Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.253841 4573 scope.go:117] "RemoveContainer" containerID="c07b04544ab56fd00ae4df96e4b96e84d6d69f79b6cf30f4c73553eb5f1edd05" Dec 03 08:42:15 crc kubenswrapper[4573]: I1203 08:42:15.253851 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57" exitCode=2 Dec 03 08:42:16 crc kubenswrapper[4573]: I1203 08:42:16.264937 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.124302 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.125607 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.126187 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.156173 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kubelet-dir\") pod \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.156752 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-var-lock\") pod \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.156817 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kube-api-access\") pod \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\" (UID: \"a39c171a-2a27-4143-b1ac-ccdcba20fbc9\") " Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.156308 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a39c171a-2a27-4143-b1ac-ccdcba20fbc9" (UID: "a39c171a-2a27-4143-b1ac-ccdcba20fbc9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.156952 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-var-lock" (OuterVolumeSpecName: "var-lock") pod "a39c171a-2a27-4143-b1ac-ccdcba20fbc9" (UID: "a39c171a-2a27-4143-b1ac-ccdcba20fbc9"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.157238 4573 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.157269 4573 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.230723 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a39c171a-2a27-4143-b1ac-ccdcba20fbc9" (UID: "a39c171a-2a27-4143-b1ac-ccdcba20fbc9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.263832 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a39c171a-2a27-4143-b1ac-ccdcba20fbc9-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.280830 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a39c171a-2a27-4143-b1ac-ccdcba20fbc9","Type":"ContainerDied","Data":"7104026b6eac6dd68fc704548dcffaaf3d513fdee78caba2400cb523e95f6fca"} Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.280914 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7104026b6eac6dd68fc704548dcffaaf3d513fdee78caba2400cb523e95f6fca" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.280992 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.313534 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.314030 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.660467 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.661956 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.662891 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.663558 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.663983 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.771870 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.771929 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.771945 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.772022 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.772030 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.772140 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.772661 4573 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.772704 4573 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.772716 4573 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.819806 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.819864 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.870555 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.871707 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.872535 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.872985 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:17 crc kubenswrapper[4573]: I1203 08:42:17.873464 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.040262 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.122924 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.122983 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.179201 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.179888 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.180237 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.180419 4573 status_manager.go:851] "Failed to get status for pod" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" pod="openshift-marketplace/redhat-operators-htmvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-htmvp\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.180707 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.290973 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.291973 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97" exitCode=0 Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.292144 4573 scope.go:117] "RemoveContainer" containerID="b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.292168 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.293084 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.293771 4573 status_manager.go:851] "Failed to get status for pod" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" pod="openshift-marketplace/redhat-operators-htmvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-htmvp\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.296401 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.296884 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.297352 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.298162 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.298470 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.298816 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.300278 4573 status_manager.go:851] "Failed to get status for pod" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" pod="openshift-marketplace/redhat-operators-htmvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-htmvp\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.300541 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.335032 4573 scope.go:117] "RemoveContainer" containerID="e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.366712 4573 scope.go:117] "RemoveContainer" containerID="eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.379809 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.380258 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.380441 4573 status_manager.go:851] "Failed to get status for pod" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" pod="openshift-marketplace/redhat-operators-htmvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-htmvp\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.380624 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.380826 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.381007 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.389558 4573 scope.go:117] "RemoveContainer" containerID="091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.393558 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.394126 4573 status_manager.go:851] "Failed to get status for pod" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" pod="openshift-marketplace/redhat-operators-htmvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-htmvp\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.394336 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.394562 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.394763 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.395013 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.414147 4573 scope.go:117] "RemoveContainer" containerID="2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.435036 4573 scope.go:117] "RemoveContainer" containerID="d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.475994 4573 scope.go:117] "RemoveContainer" containerID="b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb" Dec 03 08:42:18 crc kubenswrapper[4573]: E1203 08:42:18.477297 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\": container with ID starting with b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb not found: ID does not exist" containerID="b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.477347 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb"} err="failed to get container status \"b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\": rpc error: code = NotFound desc = could not find container \"b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb\": container with ID starting with b78f192c370ba9774785ff12f2b3dc78e8a8fb8a268fa65d52ddb56c35e800fb not found: ID does not exist" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.477379 4573 scope.go:117] "RemoveContainer" containerID="e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4" Dec 03 08:42:18 crc kubenswrapper[4573]: E1203 08:42:18.477815 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\": container with ID starting with e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4 not found: ID does not exist" containerID="e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.477863 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4"} err="failed to get container status \"e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\": rpc error: code = NotFound desc = could not find container \"e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4\": container with ID starting with e6674860806cc3b54930b5d110e66fd02bc6070ff78677ba275d13969414ecd4 not found: ID does not exist" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.477963 4573 scope.go:117] "RemoveContainer" containerID="eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9" Dec 03 08:42:18 crc kubenswrapper[4573]: E1203 08:42:18.478314 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\": container with ID starting with eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9 not found: ID does not exist" containerID="eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.478362 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9"} err="failed to get container status \"eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\": rpc error: code = NotFound desc = could not find container \"eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9\": container with ID starting with eb30210606a6a8ed6c3dcbeca0f400ffce947afea5261543f716cd6350073bc9 not found: ID does not exist" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.478386 4573 scope.go:117] "RemoveContainer" containerID="091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57" Dec 03 08:42:18 crc kubenswrapper[4573]: E1203 08:42:18.479028 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\": container with ID starting with 091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57 not found: ID does not exist" containerID="091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.479137 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57"} err="failed to get container status \"091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\": rpc error: code = NotFound desc = could not find container \"091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57\": container with ID starting with 091960eac84d5d0d5b150a2d50095d49e27bce1470daf4c16f3105d4d75d2c57 not found: ID does not exist" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.479156 4573 scope.go:117] "RemoveContainer" containerID="2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97" Dec 03 08:42:18 crc kubenswrapper[4573]: E1203 08:42:18.479473 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\": container with ID starting with 2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97 not found: ID does not exist" containerID="2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.479514 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97"} err="failed to get container status \"2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\": rpc error: code = NotFound desc = could not find container \"2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97\": container with ID starting with 2924e6a55f44e1d116a0a1f300140371b2b8a3174e01bdf498ac8bf74c9d7c97 not found: ID does not exist" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.479530 4573 scope.go:117] "RemoveContainer" containerID="d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966" Dec 03 08:42:18 crc kubenswrapper[4573]: E1203 08:42:18.479796 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\": container with ID starting with d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966 not found: ID does not exist" containerID="d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966" Dec 03 08:42:18 crc kubenswrapper[4573]: I1203 08:42:18.479842 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966"} err="failed to get container status \"d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\": rpc error: code = NotFound desc = could not find container \"d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966\": container with ID starting with d9e7b0df684990a47f20ea5b10ef9b2935afe47b72fabf5e5b26621ee3463966 not found: ID does not exist" Dec 03 08:42:19 crc kubenswrapper[4573]: E1203 08:42:19.827751 4573 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:19 crc kubenswrapper[4573]: E1203 08:42:19.827987 4573 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:19 crc kubenswrapper[4573]: E1203 08:42:19.828211 4573 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:19 crc kubenswrapper[4573]: E1203 08:42:19.828445 4573 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:19 crc kubenswrapper[4573]: E1203 08:42:19.828625 4573 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:19 crc kubenswrapper[4573]: I1203 08:42:19.828647 4573 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 08:42:19 crc kubenswrapper[4573]: E1203 08:42:19.828818 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="200ms" Dec 03 08:42:20 crc kubenswrapper[4573]: E1203 08:42:20.029512 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="400ms" Dec 03 08:42:20 crc kubenswrapper[4573]: I1203 08:42:20.043486 4573 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:20 crc kubenswrapper[4573]: I1203 08:42:20.044457 4573 status_manager.go:851] "Failed to get status for pod" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" pod="openshift-marketplace/redhat-operators-htmvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-htmvp\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:20 crc kubenswrapper[4573]: I1203 08:42:20.044779 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:20 crc kubenswrapper[4573]: I1203 08:42:20.045117 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:20 crc kubenswrapper[4573]: I1203 08:42:20.045457 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:20 crc kubenswrapper[4573]: E1203 08:42:20.432475 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="800ms" Dec 03 08:42:21 crc kubenswrapper[4573]: E1203 08:42:21.181109 4573 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187da7fd2ef38c10 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Created,Message:Created container startup-monitor,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 08:42:14.855379984 +0000 UTC m=+255.423759243,LastTimestamp:2025-12-03 08:42:14.855379984 +0000 UTC m=+255.423759243,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 08:42:21 crc kubenswrapper[4573]: E1203 08:42:21.233396 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="1.6s" Dec 03 08:42:22 crc kubenswrapper[4573]: E1203 08:42:22.834638 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="3.2s" Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.030519 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.032333 4573 status_manager.go:851] "Failed to get status for pod" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" pod="openshift-marketplace/redhat-operators-htmvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-htmvp\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.032829 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.033306 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.033789 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.050885 4573 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.051238 4573 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:25 crc kubenswrapper[4573]: E1203 08:42:25.052185 4573 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.052924 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:25 crc kubenswrapper[4573]: W1203 08:42:25.082450 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-8e49873d60cd39b63d64f961eeb6a577fd8461899262eadf2e05dcf784d68ef1 WatchSource:0}: Error finding container 8e49873d60cd39b63d64f961eeb6a577fd8461899262eadf2e05dcf784d68ef1: Status 404 returned error can't find the container with id 8e49873d60cd39b63d64f961eeb6a577fd8461899262eadf2e05dcf784d68ef1 Dec 03 08:42:25 crc kubenswrapper[4573]: I1203 08:42:25.353301 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8e49873d60cd39b63d64f961eeb6a577fd8461899262eadf2e05dcf784d68ef1"} Dec 03 08:42:26 crc kubenswrapper[4573]: E1203 08:42:26.036641 4573 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="6.4s" Dec 03 08:42:26 crc kubenswrapper[4573]: I1203 08:42:26.360212 4573 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="32658720f45b2b2077ab49ad68a5b2ef42e716d568f19cdd119cb6390cdad589" exitCode=0 Dec 03 08:42:26 crc kubenswrapper[4573]: I1203 08:42:26.360268 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"32658720f45b2b2077ab49ad68a5b2ef42e716d568f19cdd119cb6390cdad589"} Dec 03 08:42:26 crc kubenswrapper[4573]: I1203 08:42:26.360815 4573 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:26 crc kubenswrapper[4573]: I1203 08:42:26.360865 4573 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:26 crc kubenswrapper[4573]: I1203 08:42:26.361114 4573 status_manager.go:851] "Failed to get status for pod" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:26 crc kubenswrapper[4573]: I1203 08:42:26.361645 4573 status_manager.go:851] "Failed to get status for pod" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" pod="openshift-marketplace/redhat-operators-htmvp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-htmvp\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:26 crc kubenswrapper[4573]: E1203 08:42:26.361649 4573 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:26 crc kubenswrapper[4573]: I1203 08:42:26.361930 4573 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:26 crc kubenswrapper[4573]: I1203 08:42:26.362381 4573 status_manager.go:851] "Failed to get status for pod" podUID="8920781e-296c-4525-8074-99b4d9262e6a" pod="openshift-marketplace/redhat-operators-gxr7l" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-gxr7l\": dial tcp 38.102.83.251:6443: connect: connection refused" Dec 03 08:42:27 crc kubenswrapper[4573]: I1203 08:42:27.409372 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6a7252b460e72d2425f000007e94ad3a0ade3237dfae81bebe993d11ce1ad6d8"} Dec 03 08:42:27 crc kubenswrapper[4573]: I1203 08:42:27.410274 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8af871dc494d1a04bb438c35aa9d3c83495bdcaf22cd0c8164265d67bd25a639"} Dec 03 08:42:27 crc kubenswrapper[4573]: I1203 08:42:27.410289 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7ad7561292ec26c3407e2576df8e91a07d8334409c6fc7658abbef443306735d"} Dec 03 08:42:28 crc kubenswrapper[4573]: I1203 08:42:28.418098 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fffa877fc29d1bb71946c4818747f5e33d27b966f248eaeadd638b466b5f0309"} Dec 03 08:42:28 crc kubenswrapper[4573]: I1203 08:42:28.418145 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8df0d68740aeb1825071d630879a0afb8fc8a7bf31014d755f2b78816b5ed21f"} Dec 03 08:42:28 crc kubenswrapper[4573]: I1203 08:42:28.418470 4573 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:28 crc kubenswrapper[4573]: I1203 08:42:28.418490 4573 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:28 crc kubenswrapper[4573]: I1203 08:42:28.418951 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:29 crc kubenswrapper[4573]: I1203 08:42:29.426336 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 08:42:29 crc kubenswrapper[4573]: I1203 08:42:29.426625 4573 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331" exitCode=1 Dec 03 08:42:29 crc kubenswrapper[4573]: I1203 08:42:29.426672 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331"} Dec 03 08:42:29 crc kubenswrapper[4573]: I1203 08:42:29.427285 4573 scope.go:117] "RemoveContainer" containerID="74e55df1f830dbf9b28f0155bc6a46888ab66326f4a2235be47c0dffd8a34331" Dec 03 08:42:29 crc kubenswrapper[4573]: I1203 08:42:29.540831 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:42:30 crc kubenswrapper[4573]: I1203 08:42:30.053387 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:30 crc kubenswrapper[4573]: I1203 08:42:30.053793 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:30 crc kubenswrapper[4573]: I1203 08:42:30.060409 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:30 crc kubenswrapper[4573]: I1203 08:42:30.436202 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 08:42:30 crc kubenswrapper[4573]: I1203 08:42:30.436767 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"35bb452e165c5918c627ee296893d615e0fea4d4b53bece6cc61c7b7ba93fa15"} Dec 03 08:42:30 crc kubenswrapper[4573]: I1203 08:42:30.709038 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:42:30 crc kubenswrapper[4573]: I1203 08:42:30.713101 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:42:31 crc kubenswrapper[4573]: I1203 08:42:31.458574 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:42:33 crc kubenswrapper[4573]: I1203 08:42:33.515211 4573 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:33 crc kubenswrapper[4573]: I1203 08:42:33.640153 4573 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5505e97d-c389-46cc-a219-20540535b9ce" Dec 03 08:42:34 crc kubenswrapper[4573]: I1203 08:42:34.479449 4573 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:34 crc kubenswrapper[4573]: I1203 08:42:34.479496 4573 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:34 crc kubenswrapper[4573]: I1203 08:42:34.485306 4573 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5505e97d-c389-46cc-a219-20540535b9ce" Dec 03 08:42:34 crc kubenswrapper[4573]: I1203 08:42:34.485954 4573 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://7ad7561292ec26c3407e2576df8e91a07d8334409c6fc7658abbef443306735d" Dec 03 08:42:34 crc kubenswrapper[4573]: I1203 08:42:34.485988 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:35 crc kubenswrapper[4573]: I1203 08:42:35.485680 4573 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:35 crc kubenswrapper[4573]: I1203 08:42:35.485738 4573 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6679bb39-7c2c-4518-bf40-1a1abf9ebc66" Dec 03 08:42:35 crc kubenswrapper[4573]: I1203 08:42:35.489893 4573 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5505e97d-c389-46cc-a219-20540535b9ce" Dec 03 08:42:39 crc kubenswrapper[4573]: I1203 08:42:39.546035 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 08:42:40 crc kubenswrapper[4573]: I1203 08:42:40.187377 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 08:42:40 crc kubenswrapper[4573]: I1203 08:42:40.537481 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 08:42:40 crc kubenswrapper[4573]: I1203 08:42:40.669522 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 08:42:40 crc kubenswrapper[4573]: I1203 08:42:40.818503 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 08:42:40 crc kubenswrapper[4573]: I1203 08:42:40.875277 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 08:42:40 crc kubenswrapper[4573]: I1203 08:42:40.952966 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 08:42:40 crc kubenswrapper[4573]: I1203 08:42:40.965271 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 08:42:41 crc kubenswrapper[4573]: I1203 08:42:41.046238 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 08:42:41 crc kubenswrapper[4573]: I1203 08:42:41.119796 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 08:42:41 crc kubenswrapper[4573]: I1203 08:42:41.215040 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 08:42:41 crc kubenswrapper[4573]: I1203 08:42:41.597250 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 08:42:41 crc kubenswrapper[4573]: I1203 08:42:41.915350 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 08:42:41 crc kubenswrapper[4573]: I1203 08:42:41.965000 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 08:42:42 crc kubenswrapper[4573]: I1203 08:42:42.215090 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 08:42:42 crc kubenswrapper[4573]: I1203 08:42:42.347977 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 08:42:42 crc kubenswrapper[4573]: I1203 08:42:42.481946 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 08:42:42 crc kubenswrapper[4573]: I1203 08:42:42.513184 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 08:42:42 crc kubenswrapper[4573]: I1203 08:42:42.547686 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 08:42:42 crc kubenswrapper[4573]: I1203 08:42:42.846771 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 08:42:42 crc kubenswrapper[4573]: I1203 08:42:42.901772 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 08:42:43 crc kubenswrapper[4573]: I1203 08:42:43.305147 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 08:42:43 crc kubenswrapper[4573]: I1203 08:42:43.464170 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 08:42:43 crc kubenswrapper[4573]: I1203 08:42:43.514749 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 08:42:43 crc kubenswrapper[4573]: I1203 08:42:43.700420 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 08:42:43 crc kubenswrapper[4573]: I1203 08:42:43.876890 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.006794 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.075930 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.112235 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.133022 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.202396 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.367706 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.536313 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.674649 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.713503 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 08:42:44 crc kubenswrapper[4573]: I1203 08:42:44.942465 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.258444 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.523787 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.537171 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.610412 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.659600 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.670558 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.720232 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.802912 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.895784 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 08:42:45 crc kubenswrapper[4573]: I1203 08:42:45.987086 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 08:42:46 crc kubenswrapper[4573]: I1203 08:42:46.039580 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 08:42:46 crc kubenswrapper[4573]: I1203 08:42:46.152805 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 08:42:46 crc kubenswrapper[4573]: I1203 08:42:46.687793 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 08:42:47 crc kubenswrapper[4573]: I1203 08:42:47.005544 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 08:42:47 crc kubenswrapper[4573]: I1203 08:42:47.154608 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 08:42:47 crc kubenswrapper[4573]: I1203 08:42:47.179994 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 08:42:47 crc kubenswrapper[4573]: I1203 08:42:47.507773 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 08:42:47 crc kubenswrapper[4573]: I1203 08:42:47.626983 4573 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 08:42:47 crc kubenswrapper[4573]: I1203 08:42:47.829768 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 08:42:47 crc kubenswrapper[4573]: I1203 08:42:47.899876 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.242142 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.456827 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.485763 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.486539 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.551792 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.581504 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.689597 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.740084 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.767133 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.907873 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 08:42:48 crc kubenswrapper[4573]: I1203 08:42:48.971867 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.143548 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.154421 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.174737 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.178521 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.432411 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.473531 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.619571 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.716132 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.822806 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.871665 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.885653 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 08:42:49 crc kubenswrapper[4573]: I1203 08:42:49.970157 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.032487 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.277719 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.473917 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.501315 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.643777 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.673196 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.779926 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.826651 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.849708 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 08:42:50 crc kubenswrapper[4573]: I1203 08:42:50.900509 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.084188 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.127261 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.136627 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.231540 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.289180 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.325596 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.434390 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.440950 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.503592 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.520406 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.546122 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.556079 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.633242 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.710539 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.820863 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.843515 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 08:42:51 crc kubenswrapper[4573]: I1203 08:42:51.899831 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.067359 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.079146 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.086348 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.273823 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.287982 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.329415 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.463638 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.509176 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.522320 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.537354 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.582201 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.689933 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.740007 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.745701 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.822910 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.893825 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.951577 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.952037 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 08:42:52 crc kubenswrapper[4573]: I1203 08:42:52.972965 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.018715 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.033391 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.038076 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.070230 4573 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.244626 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.256002 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.389713 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.522324 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.565977 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.602964 4573 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.604453 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=39.604399745 podStartE2EDuration="39.604399745s" podCreationTimestamp="2025-12-03 08:42:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:42:33.531823332 +0000 UTC m=+274.100202591" watchObservedRunningTime="2025-12-03 08:42:53.604399745 +0000 UTC m=+294.172779004" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.606810 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.609767 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.609830 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.615898 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.643962 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.643924379 podStartE2EDuration="20.643924379s" podCreationTimestamp="2025-12-03 08:42:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:42:53.629963954 +0000 UTC m=+294.198343213" watchObservedRunningTime="2025-12-03 08:42:53.643924379 +0000 UTC m=+294.212303638" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.657791 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.746355 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.761379 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.863453 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.890638 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.993393 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 08:42:53 crc kubenswrapper[4573]: I1203 08:42:53.993523 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.095817 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.103316 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.119482 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.121207 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.140404 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.170361 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.241301 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.314947 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.353913 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.546557 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.585195 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.763780 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.943740 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.959363 4573 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.977076 4573 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 08:42:54 crc kubenswrapper[4573]: I1203 08:42:54.977350 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409" gracePeriod=5 Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.008411 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.062123 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.099757 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.188146 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.331959 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.426969 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.483576 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.561088 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.597916 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.605627 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.656260 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.658746 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.725168 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.791428 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.830405 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.833036 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.936753 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 08:42:55 crc kubenswrapper[4573]: I1203 08:42:55.996789 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.092155 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.122919 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.208260 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.352711 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.410069 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.470554 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.558696 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.659500 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.665575 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.697591 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.713103 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.749535 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.765091 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.781688 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.845418 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.910033 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 08:42:56 crc kubenswrapper[4573]: I1203 08:42:56.998851 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.042498 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.058619 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.064676 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.143735 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.150396 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.207009 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.216937 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.217929 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.254463 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.397395 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.647102 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.748131 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.868207 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.897095 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.906436 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 08:42:57 crc kubenswrapper[4573]: I1203 08:42:57.940169 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.047380 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.103452 4573 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.195353 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.403139 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.419107 4573 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.468381 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.506485 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.602170 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.644566 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5bdk6"] Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.644953 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5bdk6" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerName="registry-server" containerID="cri-o://23c53945a36baf5c3722738ca2da6593a037261aae027cd3a869f3ac2b627f87" gracePeriod=30 Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.667498 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-722wl"] Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.668484 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-722wl" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="registry-server" containerID="cri-o://750dddb0619e98e0d167fb09b34b3ce1b7fbd74f3dedcd97e57bea54d48ea67a" gracePeriod=30 Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.677993 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k2b4r"] Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.678400 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" containerID="cri-o://e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72" gracePeriod=30 Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.685656 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fzg9q"] Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.686211 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fzg9q" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerName="registry-server" containerID="cri-o://4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b" gracePeriod=30 Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.724972 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gxr7l"] Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.725479 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gxr7l" podUID="8920781e-296c-4525-8074-99b4d9262e6a" containerName="registry-server" containerID="cri-o://432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f" gracePeriod=30 Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.749663 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.779964 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-htmvp"] Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.780746 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-htmvp" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerName="registry-server" containerID="cri-o://f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e" gracePeriod=30 Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.802125 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 08:42:58 crc kubenswrapper[4573]: I1203 08:42:58.962793 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.017862 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.065249 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.124741 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.131317 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.157855 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.286497 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.475582 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.518864 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-catalog-content\") pod \"91e26815-0b3b-4855-be0a-603e52a2d02e\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.519001 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2g6r\" (UniqueName: \"kubernetes.io/projected/91e26815-0b3b-4855-be0a-603e52a2d02e-kube-api-access-t2g6r\") pod \"91e26815-0b3b-4855-be0a-603e52a2d02e\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.519136 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-utilities\") pod \"91e26815-0b3b-4855-be0a-603e52a2d02e\" (UID: \"91e26815-0b3b-4855-be0a-603e52a2d02e\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.524989 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-utilities" (OuterVolumeSpecName: "utilities") pod "91e26815-0b3b-4855-be0a-603e52a2d02e" (UID: "91e26815-0b3b-4855-be0a-603e52a2d02e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.529287 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e26815-0b3b-4855-be0a-603e52a2d02e-kube-api-access-t2g6r" (OuterVolumeSpecName: "kube-api-access-t2g6r") pod "91e26815-0b3b-4855-be0a-603e52a2d02e" (UID: "91e26815-0b3b-4855-be0a-603e52a2d02e"). InnerVolumeSpecName "kube-api-access-t2g6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.541971 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91e26815-0b3b-4855-be0a-603e52a2d02e" (UID: "91e26815-0b3b-4855-be0a-603e52a2d02e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.572963 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.577354 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.608486 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.609122 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620284 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-utilities\") pod \"8920781e-296c-4525-8074-99b4d9262e6a\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620332 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fl8d\" (UniqueName: \"kubernetes.io/projected/a44990ef-b3ba-4db7-98b2-7c6100a3acae-kube-api-access-5fl8d\") pod \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620453 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-trusted-ca\") pod \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620490 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-catalog-content\") pod \"8920781e-296c-4525-8074-99b4d9262e6a\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620517 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljrp9\" (UniqueName: \"kubernetes.io/projected/8920781e-296c-4525-8074-99b4d9262e6a-kube-api-access-ljrp9\") pod \"8920781e-296c-4525-8074-99b4d9262e6a\" (UID: \"8920781e-296c-4525-8074-99b4d9262e6a\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620540 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-operator-metrics\") pod \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\" (UID: \"a44990ef-b3ba-4db7-98b2-7c6100a3acae\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620778 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2g6r\" (UniqueName: \"kubernetes.io/projected/91e26815-0b3b-4855-be0a-603e52a2d02e-kube-api-access-t2g6r\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620793 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.620804 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e26815-0b3b-4855-be0a-603e52a2d02e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.624749 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "a44990ef-b3ba-4db7-98b2-7c6100a3acae" (UID: "a44990ef-b3ba-4db7-98b2-7c6100a3acae"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.626792 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-utilities" (OuterVolumeSpecName: "utilities") pod "8920781e-296c-4525-8074-99b4d9262e6a" (UID: "8920781e-296c-4525-8074-99b4d9262e6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.644679 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "a44990ef-b3ba-4db7-98b2-7c6100a3acae" (UID: "a44990ef-b3ba-4db7-98b2-7c6100a3acae"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.647823 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a44990ef-b3ba-4db7-98b2-7c6100a3acae-kube-api-access-5fl8d" (OuterVolumeSpecName: "kube-api-access-5fl8d") pod "a44990ef-b3ba-4db7-98b2-7c6100a3acae" (UID: "a44990ef-b3ba-4db7-98b2-7c6100a3acae"). InnerVolumeSpecName "kube-api-access-5fl8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.653658 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8920781e-296c-4525-8074-99b4d9262e6a-kube-api-access-ljrp9" (OuterVolumeSpecName: "kube-api-access-ljrp9") pod "8920781e-296c-4525-8074-99b4d9262e6a" (UID: "8920781e-296c-4525-8074-99b4d9262e6a"). InnerVolumeSpecName "kube-api-access-ljrp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.665785 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.690175 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.698508 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-722wl" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.698709 4573 generic.go:334] "Generic (PLEG): container finished" podID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerID="4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b" exitCode=0 Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.698819 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fzg9q" event={"ID":"91e26815-0b3b-4855-be0a-603e52a2d02e","Type":"ContainerDied","Data":"4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.698858 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fzg9q" event={"ID":"91e26815-0b3b-4855-be0a-603e52a2d02e","Type":"ContainerDied","Data":"2b9b4ed36bf8faf4a074ddf0b1d67fd4e193381183430c9954c7886de27feb52"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.698883 4573 scope.go:117] "RemoveContainer" containerID="4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.699029 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fzg9q" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.714648 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.726024 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6n8c\" (UniqueName: \"kubernetes.io/projected/cddd1b3a-1843-49e2-a538-051f6f698cd4-kube-api-access-q6n8c\") pod \"cddd1b3a-1843-49e2-a538-051f6f698cd4\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.726127 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-utilities\") pod \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.726246 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-catalog-content\") pod \"cddd1b3a-1843-49e2-a538-051f6f698cd4\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.726280 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95rk5\" (UniqueName: \"kubernetes.io/projected/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-kube-api-access-95rk5\") pod \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.726436 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-catalog-content\") pod \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\" (UID: \"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.726463 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-utilities\") pod \"cddd1b3a-1843-49e2-a538-051f6f698cd4\" (UID: \"cddd1b3a-1843-49e2-a538-051f6f698cd4\") " Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.731840 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-utilities" (OuterVolumeSpecName: "utilities") pod "e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" (UID: "e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.732529 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cddd1b3a-1843-49e2-a538-051f6f698cd4-kube-api-access-q6n8c" (OuterVolumeSpecName: "kube-api-access-q6n8c") pod "cddd1b3a-1843-49e2-a538-051f6f698cd4" (UID: "cddd1b3a-1843-49e2-a538-051f6f698cd4"). InnerVolumeSpecName "kube-api-access-q6n8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.734577 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-utilities" (OuterVolumeSpecName: "utilities") pod "cddd1b3a-1843-49e2-a538-051f6f698cd4" (UID: "cddd1b3a-1843-49e2-a538-051f6f698cd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.734985 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljrp9\" (UniqueName: \"kubernetes.io/projected/8920781e-296c-4525-8074-99b4d9262e6a-kube-api-access-ljrp9\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.735016 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.735035 4573 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.735072 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6n8c\" (UniqueName: \"kubernetes.io/projected/cddd1b3a-1843-49e2-a538-051f6f698cd4-kube-api-access-q6n8c\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.735089 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.735106 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.735124 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fl8d\" (UniqueName: \"kubernetes.io/projected/a44990ef-b3ba-4db7-98b2-7c6100a3acae-kube-api-access-5fl8d\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.735141 4573 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a44990ef-b3ba-4db7-98b2-7c6100a3acae-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.737354 4573 generic.go:334] "Generic (PLEG): container finished" podID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerID="750dddb0619e98e0d167fb09b34b3ce1b7fbd74f3dedcd97e57bea54d48ea67a" exitCode=0 Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.737495 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-722wl" event={"ID":"e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf","Type":"ContainerDied","Data":"750dddb0619e98e0d167fb09b34b3ce1b7fbd74f3dedcd97e57bea54d48ea67a"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.737644 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-722wl" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.739136 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.761232 4573 generic.go:334] "Generic (PLEG): container finished" podID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerID="f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e" exitCode=0 Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.761350 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htmvp" event={"ID":"cddd1b3a-1843-49e2-a538-051f6f698cd4","Type":"ContainerDied","Data":"f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.761398 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htmvp" event={"ID":"cddd1b3a-1843-49e2-a538-051f6f698cd4","Type":"ContainerDied","Data":"2bd9e52e21536b97cfe3c8c68dc5a8451dc692be86212c854461f8058d79e5ee"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.761509 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htmvp" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.770024 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-kube-api-access-95rk5" (OuterVolumeSpecName: "kube-api-access-95rk5") pod "e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" (UID: "e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf"). InnerVolumeSpecName "kube-api-access-95rk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.792669 4573 scope.go:117] "RemoveContainer" containerID="24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.801698 4573 generic.go:334] "Generic (PLEG): container finished" podID="8920781e-296c-4525-8074-99b4d9262e6a" containerID="432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f" exitCode=0 Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.801808 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxr7l" event={"ID":"8920781e-296c-4525-8074-99b4d9262e6a","Type":"ContainerDied","Data":"432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.801882 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gxr7l" event={"ID":"8920781e-296c-4525-8074-99b4d9262e6a","Type":"ContainerDied","Data":"7a497d1c01f753b02f8c0c1850fed6f0c659b05b0efc62f884947c4a49ff8c05"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.801987 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gxr7l" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.815487 4573 generic.go:334] "Generic (PLEG): container finished" podID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerID="e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72" exitCode=0 Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.815656 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" event={"ID":"a44990ef-b3ba-4db7-98b2-7c6100a3acae","Type":"ContainerDied","Data":"e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.815707 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" event={"ID":"a44990ef-b3ba-4db7-98b2-7c6100a3acae","Type":"ContainerDied","Data":"dd1f770decce4b87b293032bf73943687495835c3053f4d72e1f0c2ee6ded7c2"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.815801 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k2b4r" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.838264 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95rk5\" (UniqueName: \"kubernetes.io/projected/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-kube-api-access-95rk5\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.842596 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" (UID: "e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.857991 4573 scope.go:117] "RemoveContainer" containerID="d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.859206 4573 generic.go:334] "Generic (PLEG): container finished" podID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerID="23c53945a36baf5c3722738ca2da6593a037261aae027cd3a869f3ac2b627f87" exitCode=0 Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.859263 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bdk6" event={"ID":"bdf17da2-3e3a-4940-ae7c-680e6af80d09","Type":"ContainerDied","Data":"23c53945a36baf5c3722738ca2da6593a037261aae027cd3a869f3ac2b627f87"} Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.868187 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fzg9q"] Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.876887 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fzg9q"] Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.882607 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k2b4r"] Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.885837 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k2b4r"] Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.913506 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8920781e-296c-4525-8074-99b4d9262e6a" (UID: "8920781e-296c-4525-8074-99b4d9262e6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.952743 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8920781e-296c-4525-8074-99b4d9262e6a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.952780 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.962611 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cddd1b3a-1843-49e2-a538-051f6f698cd4" (UID: "cddd1b3a-1843-49e2-a538-051f6f698cd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.966727 4573 scope.go:117] "RemoveContainer" containerID="4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b" Dec 03 08:42:59 crc kubenswrapper[4573]: E1203 08:42:59.967246 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b\": container with ID starting with 4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b not found: ID does not exist" containerID="4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.967306 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b"} err="failed to get container status \"4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b\": rpc error: code = NotFound desc = could not find container \"4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b\": container with ID starting with 4aa99620c855a6defe0495a40c868bf0724803cf5f05986f14cd2684fc00d14b not found: ID does not exist" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.967343 4573 scope.go:117] "RemoveContainer" containerID="24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.967511 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:42:59 crc kubenswrapper[4573]: E1203 08:42:59.968711 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76\": container with ID starting with 24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76 not found: ID does not exist" containerID="24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.968747 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76"} err="failed to get container status \"24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76\": rpc error: code = NotFound desc = could not find container \"24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76\": container with ID starting with 24c534128916d3de3a28e3cc52dc1697fe80c38145f337e890a20623020a0d76 not found: ID does not exist" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.968773 4573 scope.go:117] "RemoveContainer" containerID="d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153" Dec 03 08:42:59 crc kubenswrapper[4573]: E1203 08:42:59.969176 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153\": container with ID starting with d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153 not found: ID does not exist" containerID="d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.969229 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153"} err="failed to get container status \"d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153\": rpc error: code = NotFound desc = could not find container \"d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153\": container with ID starting with d337d1ae6e6f0b52b9a6ced3ef7f1da0ab5b0afe37bc946de22b909d23d25153 not found: ID does not exist" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.969246 4573 scope.go:117] "RemoveContainer" containerID="750dddb0619e98e0d167fb09b34b3ce1b7fbd74f3dedcd97e57bea54d48ea67a" Dec 03 08:42:59 crc kubenswrapper[4573]: I1203 08:42:59.997651 4573 scope.go:117] "RemoveContainer" containerID="c3ab589892625d9410c779c47e4677108500f10d7d9213ff18e6869bab567d00" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.027797 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.033863 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.037103 4573 scope.go:117] "RemoveContainer" containerID="d409692cda2225631b89b257e2ebbea2dbcfbfe2aa714fc2cec8b93eb5e1b121" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.055295 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnqgr\" (UniqueName: \"kubernetes.io/projected/bdf17da2-3e3a-4940-ae7c-680e6af80d09-kube-api-access-pnqgr\") pod \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.056859 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-catalog-content\") pod \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.056973 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-utilities\") pod \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\" (UID: \"bdf17da2-3e3a-4940-ae7c-680e6af80d09\") " Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.061381 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-utilities" (OuterVolumeSpecName: "utilities") pod "bdf17da2-3e3a-4940-ae7c-680e6af80d09" (UID: "bdf17da2-3e3a-4940-ae7c-680e6af80d09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.063506 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf17da2-3e3a-4940-ae7c-680e6af80d09-kube-api-access-pnqgr" (OuterVolumeSpecName: "kube-api-access-pnqgr") pod "bdf17da2-3e3a-4940-ae7c-680e6af80d09" (UID: "bdf17da2-3e3a-4940-ae7c-680e6af80d09"). InnerVolumeSpecName "kube-api-access-pnqgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.063795 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cddd1b3a-1843-49e2-a538-051f6f698cd4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.063685 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" path="/var/lib/kubelet/pods/91e26815-0b3b-4855-be0a-603e52a2d02e/volumes" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.066259 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" path="/var/lib/kubelet/pods/a44990ef-b3ba-4db7-98b2-7c6100a3acae/volumes" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.067859 4573 scope.go:117] "RemoveContainer" containerID="f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.076776 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.142230 4573 scope.go:117] "RemoveContainer" containerID="83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.149362 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-722wl"] Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.158158 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-722wl"] Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.163948 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdf17da2-3e3a-4940-ae7c-680e6af80d09" (UID: "bdf17da2-3e3a-4940-ae7c-680e6af80d09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.166815 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnqgr\" (UniqueName: \"kubernetes.io/projected/bdf17da2-3e3a-4940-ae7c-680e6af80d09-kube-api-access-pnqgr\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.166845 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.166857 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdf17da2-3e3a-4940-ae7c-680e6af80d09-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.174394 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gxr7l"] Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.181729 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gxr7l"] Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.184914 4573 scope.go:117] "RemoveContainer" containerID="758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.190933 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-htmvp"] Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.196379 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-htmvp"] Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.216553 4573 scope.go:117] "RemoveContainer" containerID="f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e" Dec 03 08:43:00 crc kubenswrapper[4573]: E1203 08:43:00.217295 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e\": container with ID starting with f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e not found: ID does not exist" containerID="f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.217337 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e"} err="failed to get container status \"f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e\": rpc error: code = NotFound desc = could not find container \"f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e\": container with ID starting with f8ce7475cc76b7f3b05ea64c0a724ed84ac34f42f3a991bb4e33c2b07a7ad89e not found: ID does not exist" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.217371 4573 scope.go:117] "RemoveContainer" containerID="83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969" Dec 03 08:43:00 crc kubenswrapper[4573]: E1203 08:43:00.218481 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969\": container with ID starting with 83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969 not found: ID does not exist" containerID="83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.218642 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969"} err="failed to get container status \"83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969\": rpc error: code = NotFound desc = could not find container \"83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969\": container with ID starting with 83d1f838c0a5af9c7bc8152ea95f3aa0475eab633cc68dd57a4a0e94f5ff4969 not found: ID does not exist" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.218771 4573 scope.go:117] "RemoveContainer" containerID="758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081" Dec 03 08:43:00 crc kubenswrapper[4573]: E1203 08:43:00.219660 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081\": container with ID starting with 758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081 not found: ID does not exist" containerID="758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.219772 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081"} err="failed to get container status \"758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081\": rpc error: code = NotFound desc = could not find container \"758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081\": container with ID starting with 758ef72fc8290bc5b01c0d72d0049c4588a857562096edd1dafffd0551f05081 not found: ID does not exist" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.219867 4573 scope.go:117] "RemoveContainer" containerID="432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.235923 4573 scope.go:117] "RemoveContainer" containerID="a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.246261 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.258411 4573 scope.go:117] "RemoveContainer" containerID="8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.275478 4573 scope.go:117] "RemoveContainer" containerID="432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f" Dec 03 08:43:00 crc kubenswrapper[4573]: E1203 08:43:00.276265 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f\": container with ID starting with 432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f not found: ID does not exist" containerID="432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.276315 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f"} err="failed to get container status \"432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f\": rpc error: code = NotFound desc = could not find container \"432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f\": container with ID starting with 432deea05fe65f86bf9fb05be6141f81b1f2822a1b9c04b0e9e1156c388c7e5f not found: ID does not exist" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.276348 4573 scope.go:117] "RemoveContainer" containerID="a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d" Dec 03 08:43:00 crc kubenswrapper[4573]: E1203 08:43:00.276926 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d\": container with ID starting with a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d not found: ID does not exist" containerID="a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.276954 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d"} err="failed to get container status \"a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d\": rpc error: code = NotFound desc = could not find container \"a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d\": container with ID starting with a6fe28dcbb0bb0ad3e3f0d8225e62669fbaf5ec8a1e1c5a7c380b73283c9d55d not found: ID does not exist" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.276973 4573 scope.go:117] "RemoveContainer" containerID="8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8" Dec 03 08:43:00 crc kubenswrapper[4573]: E1203 08:43:00.277245 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8\": container with ID starting with 8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8 not found: ID does not exist" containerID="8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.277273 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8"} err="failed to get container status \"8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8\": rpc error: code = NotFound desc = could not find container \"8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8\": container with ID starting with 8acd9a0b7b0664a80c68339c0bb16181e0e7f31a9c6f02b0242fb679671467e8 not found: ID does not exist" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.277326 4573 scope.go:117] "RemoveContainer" containerID="e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.302449 4573 scope.go:117] "RemoveContainer" containerID="e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72" Dec 03 08:43:00 crc kubenswrapper[4573]: E1203 08:43:00.303405 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72\": container with ID starting with e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72 not found: ID does not exist" containerID="e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.303462 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72"} err="failed to get container status \"e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72\": rpc error: code = NotFound desc = could not find container \"e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72\": container with ID starting with e886047788d2845f3766cee769b3caa5fc95ce5e58e748aebd1a373e64cf7f72 not found: ID does not exist" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.310555 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.346064 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.548887 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.548977 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.673679 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.673738 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.673770 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.673816 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.673853 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.674422 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.674459 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.674436 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.674553 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.679927 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.775204 4573 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.775269 4573 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.775285 4573 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.775301 4573 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.775313 4573 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.867194 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.867272 4573 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409" exitCode=137 Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.867395 4573 scope.go:117] "RemoveContainer" containerID="046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.867551 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.873878 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5bdk6" event={"ID":"bdf17da2-3e3a-4940-ae7c-680e6af80d09","Type":"ContainerDied","Data":"631f98eb8469838fb3423ce4aa54d8b2c07c736f28c0c0b668d8fbd8fda1dd8a"} Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.874035 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5bdk6" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.896614 4573 scope.go:117] "RemoveContainer" containerID="046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409" Dec 03 08:43:00 crc kubenswrapper[4573]: E1203 08:43:00.898107 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409\": container with ID starting with 046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409 not found: ID does not exist" containerID="046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.898186 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409"} err="failed to get container status \"046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409\": rpc error: code = NotFound desc = could not find container \"046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409\": container with ID starting with 046f155e74af22ac67a00c23fae71d28c17d94947f7fbe70e25e86dffe09f409 not found: ID does not exist" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.898228 4573 scope.go:117] "RemoveContainer" containerID="23c53945a36baf5c3722738ca2da6593a037261aae027cd3a869f3ac2b627f87" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.914221 4573 scope.go:117] "RemoveContainer" containerID="f82d0601ad80b179fe25c350291e5435102f146c54c3ce30c7c1125edfcc7d84" Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.930824 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5bdk6"] Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.935447 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5bdk6"] Dec 03 08:43:00 crc kubenswrapper[4573]: I1203 08:43:00.951663 4573 scope.go:117] "RemoveContainer" containerID="84010d0a1dd1d90a422b835b40ba3ca6ed6719c991a72e1cfdb2b5093ea70b08" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.277933 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hgwmj"] Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278276 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278298 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278312 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278320 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278329 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278337 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278345 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" containerName="installer" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278351 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" containerName="installer" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278358 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278365 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278372 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8920781e-296c-4525-8074-99b4d9262e6a" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278378 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8920781e-296c-4525-8074-99b4d9262e6a" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278386 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278393 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278404 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278412 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278422 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278429 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278438 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278444 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278452 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8920781e-296c-4525-8074-99b4d9262e6a" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278458 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8920781e-296c-4525-8074-99b4d9262e6a" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278465 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278471 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278478 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278484 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278492 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8920781e-296c-4525-8074-99b4d9262e6a" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278498 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8920781e-296c-4525-8074-99b4d9262e6a" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278508 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278514 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerName="extract-utilities" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278526 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278534 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278542 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278548 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: E1203 08:43:01.278556 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278561 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="extract-content" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278698 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278709 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278717 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278727 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="a39c171a-2a27-4143-b1ac-ccdcba20fbc9" containerName="installer" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278736 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e26815-0b3b-4855-be0a-603e52a2d02e" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278745 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8920781e-296c-4525-8074-99b4d9262e6a" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278754 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="a44990ef-b3ba-4db7-98b2-7c6100a3acae" containerName="marketplace-operator" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.278763 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" containerName="registry-server" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.279321 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.283875 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.284222 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.284484 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.284815 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.285439 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hgwmj"] Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.292716 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.314865 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.384594 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e82d702a-82de-47e9-89fe-8ca23ce8a98b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.384665 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e82d702a-82de-47e9-89fe-8ca23ce8a98b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.384711 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvwh7\" (UniqueName: \"kubernetes.io/projected/e82d702a-82de-47e9-89fe-8ca23ce8a98b-kube-api-access-wvwh7\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.434017 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.486513 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e82d702a-82de-47e9-89fe-8ca23ce8a98b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.486578 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e82d702a-82de-47e9-89fe-8ca23ce8a98b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.486622 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvwh7\" (UniqueName: \"kubernetes.io/projected/e82d702a-82de-47e9-89fe-8ca23ce8a98b-kube-api-access-wvwh7\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.488384 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e82d702a-82de-47e9-89fe-8ca23ce8a98b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.494147 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e82d702a-82de-47e9-89fe-8ca23ce8a98b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.546374 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvwh7\" (UniqueName: \"kubernetes.io/projected/e82d702a-82de-47e9-89fe-8ca23ce8a98b-kube-api-access-wvwh7\") pod \"marketplace-operator-79b997595-hgwmj\" (UID: \"e82d702a-82de-47e9-89fe-8ca23ce8a98b\") " pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.579341 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.600856 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.817090 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hgwmj"] Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.823311 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.846221 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.912263 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 08:43:01 crc kubenswrapper[4573]: I1203 08:43:01.938630 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" event={"ID":"e82d702a-82de-47e9-89fe-8ca23ce8a98b","Type":"ContainerStarted","Data":"409bf65633926f165edf02555930d537ed321c55737833c9989fc53981c0ab95"} Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.024692 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.037987 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8920781e-296c-4525-8074-99b4d9262e6a" path="/var/lib/kubelet/pods/8920781e-296c-4525-8074-99b4d9262e6a/volumes" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.039573 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf17da2-3e3a-4940-ae7c-680e6af80d09" path="/var/lib/kubelet/pods/bdf17da2-3e3a-4940-ae7c-680e6af80d09/volumes" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.040224 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cddd1b3a-1843-49e2-a538-051f6f698cd4" path="/var/lib/kubelet/pods/cddd1b3a-1843-49e2-a538-051f6f698cd4/volumes" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.041331 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf" path="/var/lib/kubelet/pods/e31fa3f2-2ef3-4b7f-84e3-3be28ffdfdaf/volumes" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.041811 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.042257 4573 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.053585 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.053651 4573 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="3c5c2f59-0638-4d6e-b219-09534af00456" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.056894 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.056928 4573 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="3c5c2f59-0638-4d6e-b219-09534af00456" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.115264 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.136481 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.293580 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.296963 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.368962 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.393232 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.946309 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" event={"ID":"e82d702a-82de-47e9-89fe-8ca23ce8a98b","Type":"ContainerStarted","Data":"a1909321212fb1df7a95152fa33547b12cbee8d9b14bef3686a29518e6ad24ee"} Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.946646 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.954196 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" Dec 03 08:43:02 crc kubenswrapper[4573]: I1203 08:43:02.969792 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hgwmj" podStartSLOduration=4.969770803 podStartE2EDuration="4.969770803s" podCreationTimestamp="2025-12-03 08:42:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:43:02.968150229 +0000 UTC m=+303.536529498" watchObservedRunningTime="2025-12-03 08:43:02.969770803 +0000 UTC m=+303.538150072" Dec 03 08:43:03 crc kubenswrapper[4573]: I1203 08:43:03.080122 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 08:43:03 crc kubenswrapper[4573]: I1203 08:43:03.222240 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 08:43:04 crc kubenswrapper[4573]: I1203 08:43:04.291306 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.044937 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6v4g4"] Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.066523 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.070831 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.077390 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6v4g4"] Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.146470 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6wlqk"] Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.146767 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" podUID="a77aa3c0-2caf-48c1-a771-d374b1d3e5de" containerName="controller-manager" containerID="cri-o://590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3" gracePeriod=30 Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.204655 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-utilities\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.204735 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqwhk\" (UniqueName: \"kubernetes.io/projected/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-kube-api-access-nqwhk\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.204778 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-catalog-content\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.236366 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tcg9b"] Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.238246 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.244777 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.248635 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn"] Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.248875 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" podUID="ee739812-3e92-42ff-a39d-fcea79a3164c" containerName="route-controller-manager" containerID="cri-o://d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8" gracePeriod=30 Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.260988 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tcg9b"] Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.306013 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-utilities\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.306111 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-catalog-content\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.306148 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqwhk\" (UniqueName: \"kubernetes.io/projected/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-kube-api-access-nqwhk\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.306998 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-utilities\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.307246 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-catalog-content\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.334332 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqwhk\" (UniqueName: \"kubernetes.io/projected/caa010bb-5d5b-45f6-98b6-d2431c39a5fe-kube-api-access-nqwhk\") pod \"redhat-marketplace-6v4g4\" (UID: \"caa010bb-5d5b-45f6-98b6-d2431c39a5fe\") " pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.410680 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.410800 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-utilities\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.410881 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-catalog-content\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.410970 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr2gv\" (UniqueName: \"kubernetes.io/projected/09312f5f-d77a-431f-b16b-f31a9cde4335-kube-api-access-kr2gv\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.513594 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr2gv\" (UniqueName: \"kubernetes.io/projected/09312f5f-d77a-431f-b16b-f31a9cde4335-kube-api-access-kr2gv\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.513706 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-utilities\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.513749 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-catalog-content\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.514564 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-utilities\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.514551 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.515502 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-catalog-content\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.553849 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr2gv\" (UniqueName: \"kubernetes.io/projected/09312f5f-d77a-431f-b16b-f31a9cde4335-kube-api-access-kr2gv\") pod \"redhat-operators-tcg9b\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.566159 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.703999 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.716796 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-config\") pod \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.716896 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59pwc\" (UniqueName: \"kubernetes.io/projected/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-kube-api-access-59pwc\") pod \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.716970 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-serving-cert\") pod \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.717008 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-proxy-ca-bundles\") pod \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.717031 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-client-ca\") pod \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\" (UID: \"a77aa3c0-2caf-48c1-a771-d374b1d3e5de\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.718280 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-client-ca" (OuterVolumeSpecName: "client-ca") pod "a77aa3c0-2caf-48c1-a771-d374b1d3e5de" (UID: "a77aa3c0-2caf-48c1-a771-d374b1d3e5de"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.718949 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-config" (OuterVolumeSpecName: "config") pod "a77aa3c0-2caf-48c1-a771-d374b1d3e5de" (UID: "a77aa3c0-2caf-48c1-a771-d374b1d3e5de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.722566 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a77aa3c0-2caf-48c1-a771-d374b1d3e5de" (UID: "a77aa3c0-2caf-48c1-a771-d374b1d3e5de"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.727417 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a77aa3c0-2caf-48c1-a771-d374b1d3e5de" (UID: "a77aa3c0-2caf-48c1-a771-d374b1d3e5de"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.727817 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-kube-api-access-59pwc" (OuterVolumeSpecName: "kube-api-access-59pwc") pod "a77aa3c0-2caf-48c1-a771-d374b1d3e5de" (UID: "a77aa3c0-2caf-48c1-a771-d374b1d3e5de"). InnerVolumeSpecName "kube-api-access-59pwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.765764 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6v4g4"] Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818067 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee739812-3e92-42ff-a39d-fcea79a3164c-serving-cert\") pod \"ee739812-3e92-42ff-a39d-fcea79a3164c\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818160 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8p4l\" (UniqueName: \"kubernetes.io/projected/ee739812-3e92-42ff-a39d-fcea79a3164c-kube-api-access-n8p4l\") pod \"ee739812-3e92-42ff-a39d-fcea79a3164c\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818203 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-client-ca\") pod \"ee739812-3e92-42ff-a39d-fcea79a3164c\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818226 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-config\") pod \"ee739812-3e92-42ff-a39d-fcea79a3164c\" (UID: \"ee739812-3e92-42ff-a39d-fcea79a3164c\") " Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818523 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818535 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59pwc\" (UniqueName: \"kubernetes.io/projected/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-kube-api-access-59pwc\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818546 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818557 4573 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.818565 4573 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a77aa3c0-2caf-48c1-a771-d374b1d3e5de-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.820472 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-client-ca" (OuterVolumeSpecName: "client-ca") pod "ee739812-3e92-42ff-a39d-fcea79a3164c" (UID: "ee739812-3e92-42ff-a39d-fcea79a3164c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.820484 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-config" (OuterVolumeSpecName: "config") pod "ee739812-3e92-42ff-a39d-fcea79a3164c" (UID: "ee739812-3e92-42ff-a39d-fcea79a3164c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.824092 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee739812-3e92-42ff-a39d-fcea79a3164c-kube-api-access-n8p4l" (OuterVolumeSpecName: "kube-api-access-n8p4l") pod "ee739812-3e92-42ff-a39d-fcea79a3164c" (UID: "ee739812-3e92-42ff-a39d-fcea79a3164c"). InnerVolumeSpecName "kube-api-access-n8p4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.825019 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee739812-3e92-42ff-a39d-fcea79a3164c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ee739812-3e92-42ff-a39d-fcea79a3164c" (UID: "ee739812-3e92-42ff-a39d-fcea79a3164c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.920386 4573 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.920600 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee739812-3e92-42ff-a39d-fcea79a3164c-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.920629 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee739812-3e92-42ff-a39d-fcea79a3164c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:27 crc kubenswrapper[4573]: I1203 08:43:27.920644 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8p4l\" (UniqueName: \"kubernetes.io/projected/ee739812-3e92-42ff-a39d-fcea79a3164c-kube-api-access-n8p4l\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.039938 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tcg9b"] Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.132997 4573 generic.go:334] "Generic (PLEG): container finished" podID="ee739812-3e92-42ff-a39d-fcea79a3164c" containerID="d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8" exitCode=0 Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.133110 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" event={"ID":"ee739812-3e92-42ff-a39d-fcea79a3164c","Type":"ContainerDied","Data":"d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8"} Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.133178 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.133202 4573 scope.go:117] "RemoveContainer" containerID="d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.133183 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn" event={"ID":"ee739812-3e92-42ff-a39d-fcea79a3164c","Type":"ContainerDied","Data":"a7e972a7d2394f7a3753f501aeb3bd9f35263590fd328c1fee296d6c499de954"} Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.142104 4573 generic.go:334] "Generic (PLEG): container finished" podID="a77aa3c0-2caf-48c1-a771-d374b1d3e5de" containerID="590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3" exitCode=0 Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.142254 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.142279 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" event={"ID":"a77aa3c0-2caf-48c1-a771-d374b1d3e5de","Type":"ContainerDied","Data":"590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3"} Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.143103 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6wlqk" event={"ID":"a77aa3c0-2caf-48c1-a771-d374b1d3e5de","Type":"ContainerDied","Data":"8b8415d7dbdb09ecbab0ec26b0cba456aee7e293bbe4b381d525a1d79ad19a60"} Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.147805 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6v4g4" event={"ID":"caa010bb-5d5b-45f6-98b6-d2431c39a5fe","Type":"ContainerDied","Data":"6b47b9bf5bc21829b553567eb5e6b14888c62de1e9315a33399e42d405b81724"} Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.147634 4573 generic.go:334] "Generic (PLEG): container finished" podID="caa010bb-5d5b-45f6-98b6-d2431c39a5fe" containerID="6b47b9bf5bc21829b553567eb5e6b14888c62de1e9315a33399e42d405b81724" exitCode=0 Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.148681 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6v4g4" event={"ID":"caa010bb-5d5b-45f6-98b6-d2431c39a5fe","Type":"ContainerStarted","Data":"96b5f86ef5d30b2a0ce3e1a273c1d58e8d0e97e6bae069429ffeb23092a6233c"} Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.157107 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tcg9b" event={"ID":"09312f5f-d77a-431f-b16b-f31a9cde4335","Type":"ContainerStarted","Data":"4282c6de181ce0b9211c01e994b3328d7b9742d36706666222f08050b835f1ff"} Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.158920 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn"] Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.161489 4573 scope.go:117] "RemoveContainer" containerID="d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8" Dec 03 08:43:28 crc kubenswrapper[4573]: E1203 08:43:28.162223 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8\": container with ID starting with d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8 not found: ID does not exist" containerID="d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.162289 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8"} err="failed to get container status \"d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8\": rpc error: code = NotFound desc = could not find container \"d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8\": container with ID starting with d492cf7fe4cf4d611ec65e2fae709f0361cbb5b3186a12455470486e19fdb8b8 not found: ID does not exist" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.162322 4573 scope.go:117] "RemoveContainer" containerID="590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.168451 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-bzmrn"] Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.173140 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6wlqk"] Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.177114 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6wlqk"] Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.186889 4573 scope.go:117] "RemoveContainer" containerID="590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3" Dec 03 08:43:28 crc kubenswrapper[4573]: E1203 08:43:28.188671 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3\": container with ID starting with 590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3 not found: ID does not exist" containerID="590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.188722 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3"} err="failed to get container status \"590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3\": rpc error: code = NotFound desc = could not find container \"590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3\": container with ID starting with 590deba66800c3ed714d8f85167c244f33c9a17efe1679bb28098d27953723d3 not found: ID does not exist" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.409303 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-2hkgx"] Dec 03 08:43:28 crc kubenswrapper[4573]: E1203 08:43:28.409649 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77aa3c0-2caf-48c1-a771-d374b1d3e5de" containerName="controller-manager" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.409673 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77aa3c0-2caf-48c1-a771-d374b1d3e5de" containerName="controller-manager" Dec 03 08:43:28 crc kubenswrapper[4573]: E1203 08:43:28.409695 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee739812-3e92-42ff-a39d-fcea79a3164c" containerName="route-controller-manager" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.409707 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee739812-3e92-42ff-a39d-fcea79a3164c" containerName="route-controller-manager" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.409825 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee739812-3e92-42ff-a39d-fcea79a3164c" containerName="route-controller-manager" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.409849 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="a77aa3c0-2caf-48c1-a771-d374b1d3e5de" containerName="controller-manager" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.410446 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.412423 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.413471 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx"] Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.414317 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.414589 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.414667 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.414804 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.414954 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.416689 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.417160 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.418269 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.418472 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.418582 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.418488 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.419089 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.423482 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.429837 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-2hkgx"] Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.433203 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx"] Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530605 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-serving-cert\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530668 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-config\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530780 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpjq2\" (UniqueName: \"kubernetes.io/projected/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-kube-api-access-mpjq2\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530816 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqz26\" (UniqueName: \"kubernetes.io/projected/0598983a-e8b6-4a1a-8de2-006d916f5711-kube-api-access-xqz26\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530833 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-config\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530859 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-client-ca\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530876 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0598983a-e8b6-4a1a-8de2-006d916f5711-serving-cert\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530893 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-proxy-ca-bundles\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.530912 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-client-ca\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.632896 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqz26\" (UniqueName: \"kubernetes.io/projected/0598983a-e8b6-4a1a-8de2-006d916f5711-kube-api-access-xqz26\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.633425 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-config\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.633586 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-client-ca\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.633838 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0598983a-e8b6-4a1a-8de2-006d916f5711-serving-cert\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.633959 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-proxy-ca-bundles\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.634091 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-client-ca\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.634322 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-serving-cert\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.634603 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-config\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.635281 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpjq2\" (UniqueName: \"kubernetes.io/projected/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-kube-api-access-mpjq2\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.635011 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-client-ca\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.634895 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-config\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.635623 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-client-ca\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.635731 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-proxy-ca-bundles\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.636869 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-config\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.638409 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-serving-cert\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.639265 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0598983a-e8b6-4a1a-8de2-006d916f5711-serving-cert\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.657023 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqz26\" (UniqueName: \"kubernetes.io/projected/0598983a-e8b6-4a1a-8de2-006d916f5711-kube-api-access-xqz26\") pod \"route-controller-manager-dbbf5946d-56dlx\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.664347 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpjq2\" (UniqueName: \"kubernetes.io/projected/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-kube-api-access-mpjq2\") pod \"controller-manager-58f8979484-2hkgx\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.742696 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:28 crc kubenswrapper[4573]: I1203 08:43:28.757825 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.084620 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-2hkgx"] Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.095670 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx"] Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.167916 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6v4g4" event={"ID":"caa010bb-5d5b-45f6-98b6-d2431c39a5fe","Type":"ContainerStarted","Data":"b81f722aa327908360707d2e907c856009441df7e0dbf06282756506fd076b71"} Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.169822 4573 generic.go:334] "Generic (PLEG): container finished" podID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerID="b38d21b10f8431cfb8eb6c66a01924317e1c93f07db424d53d81d5380932e085" exitCode=0 Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.169872 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tcg9b" event={"ID":"09312f5f-d77a-431f-b16b-f31a9cde4335","Type":"ContainerDied","Data":"b38d21b10f8431cfb8eb6c66a01924317e1c93f07db424d53d81d5380932e085"} Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.257466 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx"] Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.335214 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-2hkgx"] Dec 03 08:43:29 crc kubenswrapper[4573]: W1203 08:43:29.368030 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc00ced7f_1a8d_4ec5_93b0_23e6f7df0f63.slice/crio-56ca275a84acf1bad4194d344ea6e075f00f15ac664b4113594267f06dd585cb WatchSource:0}: Error finding container 56ca275a84acf1bad4194d344ea6e075f00f15ac664b4113594267f06dd585cb: Status 404 returned error can't find the container with id 56ca275a84acf1bad4194d344ea6e075f00f15ac664b4113594267f06dd585cb Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.457625 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6zq2f"] Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.463235 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.466315 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.488604 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6zq2f"] Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.556906 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-catalog-content\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.556967 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-utilities\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.556999 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncrmj\" (UniqueName: \"kubernetes.io/projected/650903df-cc41-472a-86e1-9c890b4a135c-kube-api-access-ncrmj\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.658228 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-catalog-content\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.658282 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-utilities\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.658314 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncrmj\" (UniqueName: \"kubernetes.io/projected/650903df-cc41-472a-86e1-9c890b4a135c-kube-api-access-ncrmj\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.659078 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-catalog-content\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.659323 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-utilities\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.663867 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5rtb9"] Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.667440 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.670290 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.689187 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncrmj\" (UniqueName: \"kubernetes.io/projected/650903df-cc41-472a-86e1-9c890b4a135c-kube-api-access-ncrmj\") pod \"certified-operators-6zq2f\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.699954 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rtb9"] Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.760443 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/577aa218-59b2-4cbc-ae72-59c607006fb5-utilities\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.760592 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/577aa218-59b2-4cbc-ae72-59c607006fb5-catalog-content\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.760632 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9q8b\" (UniqueName: \"kubernetes.io/projected/577aa218-59b2-4cbc-ae72-59c607006fb5-kube-api-access-p9q8b\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.799333 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.862034 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/577aa218-59b2-4cbc-ae72-59c607006fb5-utilities\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.862147 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/577aa218-59b2-4cbc-ae72-59c607006fb5-catalog-content\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.862188 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9q8b\" (UniqueName: \"kubernetes.io/projected/577aa218-59b2-4cbc-ae72-59c607006fb5-kube-api-access-p9q8b\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.862796 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/577aa218-59b2-4cbc-ae72-59c607006fb5-catalog-content\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.863193 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/577aa218-59b2-4cbc-ae72-59c607006fb5-utilities\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.881825 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9q8b\" (UniqueName: \"kubernetes.io/projected/577aa218-59b2-4cbc-ae72-59c607006fb5-kube-api-access-p9q8b\") pod \"community-operators-5rtb9\" (UID: \"577aa218-59b2-4cbc-ae72-59c607006fb5\") " pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:29 crc kubenswrapper[4573]: I1203 08:43:29.989919 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.069480 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a77aa3c0-2caf-48c1-a771-d374b1d3e5de" path="/var/lib/kubelet/pods/a77aa3c0-2caf-48c1-a771-d374b1d3e5de/volumes" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.071458 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee739812-3e92-42ff-a39d-fcea79a3164c" path="/var/lib/kubelet/pods/ee739812-3e92-42ff-a39d-fcea79a3164c/volumes" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.204953 4573 generic.go:334] "Generic (PLEG): container finished" podID="caa010bb-5d5b-45f6-98b6-d2431c39a5fe" containerID="b81f722aa327908360707d2e907c856009441df7e0dbf06282756506fd076b71" exitCode=0 Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.205100 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6v4g4" event={"ID":"caa010bb-5d5b-45f6-98b6-d2431c39a5fe","Type":"ContainerDied","Data":"b81f722aa327908360707d2e907c856009441df7e0dbf06282756506fd076b71"} Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.226445 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" event={"ID":"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63","Type":"ContainerStarted","Data":"7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d"} Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.226509 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" event={"ID":"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63","Type":"ContainerStarted","Data":"56ca275a84acf1bad4194d344ea6e075f00f15ac664b4113594267f06dd585cb"} Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.226647 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" podUID="c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" containerName="controller-manager" containerID="cri-o://7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d" gracePeriod=30 Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.227243 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.238717 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" podUID="0598983a-e8b6-4a1a-8de2-006d916f5711" containerName="route-controller-manager" containerID="cri-o://953245c1181bc5274134fb36b6373e746b6a838fb252621d08f74a5f3243f8ca" gracePeriod=30 Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.238988 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" event={"ID":"0598983a-e8b6-4a1a-8de2-006d916f5711","Type":"ContainerStarted","Data":"953245c1181bc5274134fb36b6373e746b6a838fb252621d08f74a5f3243f8ca"} Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.239018 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" event={"ID":"0598983a-e8b6-4a1a-8de2-006d916f5711","Type":"ContainerStarted","Data":"9a13d999333945a66df6c81cd3e4daab630db9909fb30c47b4ba5f6723b2d1b3"} Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.239301 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.246130 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.255398 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tcg9b" event={"ID":"09312f5f-d77a-431f-b16b-f31a9cde4335","Type":"ContainerStarted","Data":"aaf385ffabe1b2acf8d5cce355ea2c97b0b2d07a0d2e2a31a2fb2124399c8bcb"} Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.290680 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" podStartSLOduration=3.290650859 podStartE2EDuration="3.290650859s" podCreationTimestamp="2025-12-03 08:43:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:43:30.288501029 +0000 UTC m=+330.856880298" watchObservedRunningTime="2025-12-03 08:43:30.290650859 +0000 UTC m=+330.859030118" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.371267 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" podStartSLOduration=3.371234638 podStartE2EDuration="3.371234638s" podCreationTimestamp="2025-12-03 08:43:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:43:30.360409829 +0000 UTC m=+330.928789088" watchObservedRunningTime="2025-12-03 08:43:30.371234638 +0000 UTC m=+330.939613897" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.375913 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6zq2f"] Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.508898 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rtb9"] Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.810745 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.853617 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-99456b778-8tw6l"] Dec 03 08:43:30 crc kubenswrapper[4573]: E1203 08:43:30.853938 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" containerName="controller-manager" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.853958 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" containerName="controller-manager" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.858519 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" containerName="controller-manager" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.859323 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.869970 4573 patch_prober.go:28] interesting pod/route-controller-manager-dbbf5946d-56dlx container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": read tcp 10.217.0.2:49794->10.217.0.62:8443: read: connection reset by peer" start-of-body= Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.870024 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" podUID="0598983a-e8b6-4a1a-8de2-006d916f5711" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": read tcp 10.217.0.2:49794->10.217.0.62:8443: read: connection reset by peer" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.893864 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpjq2\" (UniqueName: \"kubernetes.io/projected/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-kube-api-access-mpjq2\") pod \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.893952 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-proxy-ca-bundles\") pod \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.894029 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-config\") pod \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.894099 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-client-ca\") pod \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.894156 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-serving-cert\") pod \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\" (UID: \"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63\") " Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.894421 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrdk9\" (UniqueName: \"kubernetes.io/projected/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-kube-api-access-nrdk9\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.894475 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-proxy-ca-bundles\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.894508 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-client-ca\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.894535 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-serving-cert\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.894585 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-config\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.896412 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-config" (OuterVolumeSpecName: "config") pod "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" (UID: "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.898367 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-client-ca" (OuterVolumeSpecName: "client-ca") pod "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" (UID: "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.899627 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" (UID: "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.899954 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-99456b778-8tw6l"] Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.903549 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-kube-api-access-mpjq2" (OuterVolumeSpecName: "kube-api-access-mpjq2") pod "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" (UID: "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63"). InnerVolumeSpecName "kube-api-access-mpjq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.910242 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" (UID: "c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.995762 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrdk9\" (UniqueName: \"kubernetes.io/projected/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-kube-api-access-nrdk9\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996105 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-proxy-ca-bundles\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996131 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-client-ca\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996157 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-serving-cert\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996206 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-config\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996248 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996260 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpjq2\" (UniqueName: \"kubernetes.io/projected/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-kube-api-access-mpjq2\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996269 4573 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996278 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.996287 4573 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.997718 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-config\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.998288 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-client-ca\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:30 crc kubenswrapper[4573]: I1203 08:43:30.999714 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-proxy-ca-bundles\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.004359 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-serving-cert\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.019350 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrdk9\" (UniqueName: \"kubernetes.io/projected/84a8db50-4c2a-40eb-ad99-1e19c9cc4455-kube-api-access-nrdk9\") pod \"controller-manager-99456b778-8tw6l\" (UID: \"84a8db50-4c2a-40eb-ad99-1e19c9cc4455\") " pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.178806 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.269954 4573 generic.go:334] "Generic (PLEG): container finished" podID="577aa218-59b2-4cbc-ae72-59c607006fb5" containerID="4119435bbb50255d217d897fd06cfe648ea0ddc958e4d3f698eaf17223f6a814" exitCode=0 Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.270020 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rtb9" event={"ID":"577aa218-59b2-4cbc-ae72-59c607006fb5","Type":"ContainerDied","Data":"4119435bbb50255d217d897fd06cfe648ea0ddc958e4d3f698eaf17223f6a814"} Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.270078 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rtb9" event={"ID":"577aa218-59b2-4cbc-ae72-59c607006fb5","Type":"ContainerStarted","Data":"8db62327428c39faab4a232c602dfe7e5e0739f418723f6bd187331c26030366"} Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.275317 4573 generic.go:334] "Generic (PLEG): container finished" podID="650903df-cc41-472a-86e1-9c890b4a135c" containerID="8e1d4e2a61b2d6a46ef5ebab4e50b91365768270f1547541cdad5e5bd7503afd" exitCode=0 Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.275426 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zq2f" event={"ID":"650903df-cc41-472a-86e1-9c890b4a135c","Type":"ContainerDied","Data":"8e1d4e2a61b2d6a46ef5ebab4e50b91365768270f1547541cdad5e5bd7503afd"} Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.275454 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zq2f" event={"ID":"650903df-cc41-472a-86e1-9c890b4a135c","Type":"ContainerStarted","Data":"298f1ff2b071f9b11c13528d91dc59b4bfd69cc44227cf7e48a1dc147d4cf501"} Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.287354 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6v4g4" event={"ID":"caa010bb-5d5b-45f6-98b6-d2431c39a5fe","Type":"ContainerStarted","Data":"273cba5bb7636308f73b74b6ea038a89ba723e209ed3999825c877c3028d2930"} Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.290332 4573 generic.go:334] "Generic (PLEG): container finished" podID="c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" containerID="7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d" exitCode=0 Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.290463 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.291522 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" event={"ID":"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63","Type":"ContainerDied","Data":"7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d"} Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.291566 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58f8979484-2hkgx" event={"ID":"c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63","Type":"ContainerDied","Data":"56ca275a84acf1bad4194d344ea6e075f00f15ac664b4113594267f06dd585cb"} Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.291585 4573 scope.go:117] "RemoveContainer" containerID="7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.298168 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-dbbf5946d-56dlx_0598983a-e8b6-4a1a-8de2-006d916f5711/route-controller-manager/0.log" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.298210 4573 generic.go:334] "Generic (PLEG): container finished" podID="0598983a-e8b6-4a1a-8de2-006d916f5711" containerID="953245c1181bc5274134fb36b6373e746b6a838fb252621d08f74a5f3243f8ca" exitCode=255 Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.299634 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" event={"ID":"0598983a-e8b6-4a1a-8de2-006d916f5711","Type":"ContainerDied","Data":"953245c1181bc5274134fb36b6373e746b6a838fb252621d08f74a5f3243f8ca"} Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.335399 4573 scope.go:117] "RemoveContainer" containerID="7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.335938 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6v4g4" podStartSLOduration=1.830479474 podStartE2EDuration="4.335907592s" podCreationTimestamp="2025-12-03 08:43:27 +0000 UTC" firstStartedPulling="2025-12-03 08:43:28.159783415 +0000 UTC m=+328.728162674" lastFinishedPulling="2025-12-03 08:43:30.665211533 +0000 UTC m=+331.233590792" observedRunningTime="2025-12-03 08:43:31.332412865 +0000 UTC m=+331.900792134" watchObservedRunningTime="2025-12-03 08:43:31.335907592 +0000 UTC m=+331.904286851" Dec 03 08:43:31 crc kubenswrapper[4573]: E1203 08:43:31.336691 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d\": container with ID starting with 7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d not found: ID does not exist" containerID="7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.336775 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d"} err="failed to get container status \"7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d\": rpc error: code = NotFound desc = could not find container \"7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d\": container with ID starting with 7779b8ff574ebf11937514dd3fcd0eb19d6d5166b993652f06f855936bfe3c8d not found: ID does not exist" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.392565 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-2hkgx"] Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.395415 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-2hkgx"] Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.459391 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-99456b778-8tw6l"] Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.728401 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-dbbf5946d-56dlx_0598983a-e8b6-4a1a-8de2-006d916f5711/route-controller-manager/0.log" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.728863 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.808882 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqz26\" (UniqueName: \"kubernetes.io/projected/0598983a-e8b6-4a1a-8de2-006d916f5711-kube-api-access-xqz26\") pod \"0598983a-e8b6-4a1a-8de2-006d916f5711\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.809210 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-client-ca\") pod \"0598983a-e8b6-4a1a-8de2-006d916f5711\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.809305 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-config\") pod \"0598983a-e8b6-4a1a-8de2-006d916f5711\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.809439 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0598983a-e8b6-4a1a-8de2-006d916f5711-serving-cert\") pod \"0598983a-e8b6-4a1a-8de2-006d916f5711\" (UID: \"0598983a-e8b6-4a1a-8de2-006d916f5711\") " Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.811565 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-client-ca" (OuterVolumeSpecName: "client-ca") pod "0598983a-e8b6-4a1a-8de2-006d916f5711" (UID: "0598983a-e8b6-4a1a-8de2-006d916f5711"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.812091 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-config" (OuterVolumeSpecName: "config") pod "0598983a-e8b6-4a1a-8de2-006d916f5711" (UID: "0598983a-e8b6-4a1a-8de2-006d916f5711"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.818394 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0598983a-e8b6-4a1a-8de2-006d916f5711-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0598983a-e8b6-4a1a-8de2-006d916f5711" (UID: "0598983a-e8b6-4a1a-8de2-006d916f5711"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.819133 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0598983a-e8b6-4a1a-8de2-006d916f5711-kube-api-access-xqz26" (OuterVolumeSpecName: "kube-api-access-xqz26") pod "0598983a-e8b6-4a1a-8de2-006d916f5711" (UID: "0598983a-e8b6-4a1a-8de2-006d916f5711"). InnerVolumeSpecName "kube-api-access-xqz26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.910542 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqz26\" (UniqueName: \"kubernetes.io/projected/0598983a-e8b6-4a1a-8de2-006d916f5711-kube-api-access-xqz26\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.910777 4573 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.910879 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0598983a-e8b6-4a1a-8de2-006d916f5711-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:31 crc kubenswrapper[4573]: I1203 08:43:31.910950 4573 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0598983a-e8b6-4a1a-8de2-006d916f5711-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.037691 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63" path="/var/lib/kubelet/pods/c00ced7f-1a8d-4ec5-93b0-23e6f7df0f63/volumes" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.293737 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45"] Dec 03 08:43:32 crc kubenswrapper[4573]: E1203 08:43:32.293963 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0598983a-e8b6-4a1a-8de2-006d916f5711" containerName="route-controller-manager" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.293974 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="0598983a-e8b6-4a1a-8de2-006d916f5711" containerName="route-controller-manager" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.294094 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="0598983a-e8b6-4a1a-8de2-006d916f5711" containerName="route-controller-manager" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.294442 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.307030 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zq2f" event={"ID":"650903df-cc41-472a-86e1-9c890b4a135c","Type":"ContainerStarted","Data":"80ff6a52d9927751b873cc37b8fe346eaec45ce6978fa671044aef9fe6bd928a"} Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.312643 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-dbbf5946d-56dlx_0598983a-e8b6-4a1a-8de2-006d916f5711/route-controller-manager/0.log" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.312793 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.312955 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx" event={"ID":"0598983a-e8b6-4a1a-8de2-006d916f5711","Type":"ContainerDied","Data":"9a13d999333945a66df6c81cd3e4daab630db9909fb30c47b4ba5f6723b2d1b3"} Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.313067 4573 scope.go:117] "RemoveContainer" containerID="953245c1181bc5274134fb36b6373e746b6a838fb252621d08f74a5f3243f8ca" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.315748 4573 generic.go:334] "Generic (PLEG): container finished" podID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerID="aaf385ffabe1b2acf8d5cce355ea2c97b0b2d07a0d2e2a31a2fb2124399c8bcb" exitCode=0 Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.315828 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tcg9b" event={"ID":"09312f5f-d77a-431f-b16b-f31a9cde4335","Type":"ContainerDied","Data":"aaf385ffabe1b2acf8d5cce355ea2c97b0b2d07a0d2e2a31a2fb2124399c8bcb"} Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.324755 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" event={"ID":"84a8db50-4c2a-40eb-ad99-1e19c9cc4455","Type":"ContainerStarted","Data":"c04d25ddbb4271a9dae566b61b1e5872c61592e9125250e7b372f883c565b2e5"} Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.324812 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" event={"ID":"84a8db50-4c2a-40eb-ad99-1e19c9cc4455","Type":"ContainerStarted","Data":"ff99e7a6cef88f3bf250c6282874fc6107a1781408c23cbc4cbf275dfafdfe6c"} Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.324831 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.334411 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45"] Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.335382 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.364106 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx"] Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.383742 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-dbbf5946d-56dlx"] Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.421176 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-client-ca\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.421308 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-config\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.421341 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-serving-cert\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.421398 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm4nw\" (UniqueName: \"kubernetes.io/projected/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-kube-api-access-wm4nw\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.462013 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" podStartSLOduration=3.461987852 podStartE2EDuration="3.461987852s" podCreationTimestamp="2025-12-03 08:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:43:32.444496128 +0000 UTC m=+333.012875387" watchObservedRunningTime="2025-12-03 08:43:32.461987852 +0000 UTC m=+333.030367111" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.522227 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-config\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.522277 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-serving-cert\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.522310 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm4nw\" (UniqueName: \"kubernetes.io/projected/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-kube-api-access-wm4nw\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.522355 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-client-ca\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.523355 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-client-ca\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.523764 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-config\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.533941 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-serving-cert\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.558410 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm4nw\" (UniqueName: \"kubernetes.io/projected/6cfe1e30-6f38-4905-94e6-a7b34b75e1a8-kube-api-access-wm4nw\") pod \"route-controller-manager-7fbf4d8f5d-2gp45\" (UID: \"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8\") " pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.617326 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:32 crc kubenswrapper[4573]: I1203 08:43:32.918288 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45"] Dec 03 08:43:32 crc kubenswrapper[4573]: W1203 08:43:32.927460 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6cfe1e30_6f38_4905_94e6_a7b34b75e1a8.slice/crio-9572fe31fc6d3c025510ec797da2a68a876ab00fa18180bcacb591208335f019 WatchSource:0}: Error finding container 9572fe31fc6d3c025510ec797da2a68a876ab00fa18180bcacb591208335f019: Status 404 returned error can't find the container with id 9572fe31fc6d3c025510ec797da2a68a876ab00fa18180bcacb591208335f019 Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.331922 4573 generic.go:334] "Generic (PLEG): container finished" podID="577aa218-59b2-4cbc-ae72-59c607006fb5" containerID="d50f853f2b563634e6927c8913c271c3a9dca674fe6608f451bf2b8577c38d3f" exitCode=0 Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.332067 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rtb9" event={"ID":"577aa218-59b2-4cbc-ae72-59c607006fb5","Type":"ContainerDied","Data":"d50f853f2b563634e6927c8913c271c3a9dca674fe6608f451bf2b8577c38d3f"} Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.336552 4573 generic.go:334] "Generic (PLEG): container finished" podID="650903df-cc41-472a-86e1-9c890b4a135c" containerID="80ff6a52d9927751b873cc37b8fe346eaec45ce6978fa671044aef9fe6bd928a" exitCode=0 Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.336624 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zq2f" event={"ID":"650903df-cc41-472a-86e1-9c890b4a135c","Type":"ContainerDied","Data":"80ff6a52d9927751b873cc37b8fe346eaec45ce6978fa671044aef9fe6bd928a"} Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.342434 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" event={"ID":"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8","Type":"ContainerStarted","Data":"54768957ccd4578310ed3b208f268c6bebea3101633ab5a7ac1e0e1bb66e3d4c"} Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.342468 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" event={"ID":"6cfe1e30-6f38-4905-94e6-a7b34b75e1a8","Type":"ContainerStarted","Data":"9572fe31fc6d3c025510ec797da2a68a876ab00fa18180bcacb591208335f019"} Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.342916 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.347583 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tcg9b" event={"ID":"09312f5f-d77a-431f-b16b-f31a9cde4335","Type":"ContainerStarted","Data":"c409c85639c839576ef225c3b40d7c68e9824fe4ebf7d8c064837b0bbe4a17b7"} Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.409121 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tcg9b" podStartSLOduration=2.861329619 podStartE2EDuration="6.40908596s" podCreationTimestamp="2025-12-03 08:43:27 +0000 UTC" firstStartedPulling="2025-12-03 08:43:29.17512104 +0000 UTC m=+329.743500299" lastFinishedPulling="2025-12-03 08:43:32.722877371 +0000 UTC m=+333.291256640" observedRunningTime="2025-12-03 08:43:33.401471119 +0000 UTC m=+333.969850388" watchObservedRunningTime="2025-12-03 08:43:33.40908596 +0000 UTC m=+333.977465219" Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.422930 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" podStartSLOduration=1.4229076520000001 podStartE2EDuration="1.422907652s" podCreationTimestamp="2025-12-03 08:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:43:33.4228327 +0000 UTC m=+333.991211969" watchObservedRunningTime="2025-12-03 08:43:33.422907652 +0000 UTC m=+333.991286911" Dec 03 08:43:33 crc kubenswrapper[4573]: I1203 08:43:33.552533 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7fbf4d8f5d-2gp45" Dec 03 08:43:34 crc kubenswrapper[4573]: I1203 08:43:34.040082 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0598983a-e8b6-4a1a-8de2-006d916f5711" path="/var/lib/kubelet/pods/0598983a-e8b6-4a1a-8de2-006d916f5711/volumes" Dec 03 08:43:34 crc kubenswrapper[4573]: I1203 08:43:34.354444 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rtb9" event={"ID":"577aa218-59b2-4cbc-ae72-59c607006fb5","Type":"ContainerStarted","Data":"c7635ed7abde117fef92a7cad40b750cf22c110eecaa828881dee708ae76e980"} Dec 03 08:43:34 crc kubenswrapper[4573]: I1203 08:43:34.358260 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zq2f" event={"ID":"650903df-cc41-472a-86e1-9c890b4a135c","Type":"ContainerStarted","Data":"3b2ee87793e9ed98386658a8263f893e4713acb7d84c761c68b146b9ea636b25"} Dec 03 08:43:34 crc kubenswrapper[4573]: I1203 08:43:34.404773 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5rtb9" podStartSLOduration=2.95017891 podStartE2EDuration="5.404757131s" podCreationTimestamp="2025-12-03 08:43:29 +0000 UTC" firstStartedPulling="2025-12-03 08:43:31.27185805 +0000 UTC m=+331.840237309" lastFinishedPulling="2025-12-03 08:43:33.726436271 +0000 UTC m=+334.294815530" observedRunningTime="2025-12-03 08:43:34.384025907 +0000 UTC m=+334.952405166" watchObservedRunningTime="2025-12-03 08:43:34.404757131 +0000 UTC m=+334.973136390" Dec 03 08:43:37 crc kubenswrapper[4573]: I1203 08:43:37.411330 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:37 crc kubenswrapper[4573]: I1203 08:43:37.411849 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:37 crc kubenswrapper[4573]: I1203 08:43:37.470680 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:37 crc kubenswrapper[4573]: I1203 08:43:37.496429 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6zq2f" podStartSLOduration=6.032767488 podStartE2EDuration="8.49640173s" podCreationTimestamp="2025-12-03 08:43:29 +0000 UTC" firstStartedPulling="2025-12-03 08:43:31.277188667 +0000 UTC m=+331.845567916" lastFinishedPulling="2025-12-03 08:43:33.740822899 +0000 UTC m=+334.309202158" observedRunningTime="2025-12-03 08:43:34.408779932 +0000 UTC m=+334.977159201" watchObservedRunningTime="2025-12-03 08:43:37.49640173 +0000 UTC m=+338.064780989" Dec 03 08:43:37 crc kubenswrapper[4573]: I1203 08:43:37.567688 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:37 crc kubenswrapper[4573]: I1203 08:43:37.568133 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:38 crc kubenswrapper[4573]: I1203 08:43:38.446485 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6v4g4" Dec 03 08:43:38 crc kubenswrapper[4573]: I1203 08:43:38.614294 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tcg9b" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="registry-server" probeResult="failure" output=< Dec 03 08:43:38 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 08:43:38 crc kubenswrapper[4573]: > Dec 03 08:43:39 crc kubenswrapper[4573]: I1203 08:43:39.799986 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:39 crc kubenswrapper[4573]: I1203 08:43:39.800524 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:39 crc kubenswrapper[4573]: I1203 08:43:39.860574 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:39 crc kubenswrapper[4573]: I1203 08:43:39.990787 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:39 crc kubenswrapper[4573]: I1203 08:43:39.990922 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:40 crc kubenswrapper[4573]: I1203 08:43:40.039536 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:40 crc kubenswrapper[4573]: I1203 08:43:40.499804 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 08:43:40 crc kubenswrapper[4573]: I1203 08:43:40.508574 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5rtb9" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.485207 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ckl57"] Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.487884 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.503191 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ckl57"] Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.582025 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d82edf84-5eee-4dab-b5f7-459a986a8637-trusted-ca\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.582375 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d82edf84-5eee-4dab-b5f7-459a986a8637-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.582485 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d82edf84-5eee-4dab-b5f7-459a986a8637-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.582597 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d82edf84-5eee-4dab-b5f7-459a986a8637-registry-certificates\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.582713 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzrpf\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-kube-api-access-bzrpf\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.582866 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.582990 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-registry-tls\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.583149 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-bound-sa-token\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.620739 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.684838 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d82edf84-5eee-4dab-b5f7-459a986a8637-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.685225 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d82edf84-5eee-4dab-b5f7-459a986a8637-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.685334 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d82edf84-5eee-4dab-b5f7-459a986a8637-registry-certificates\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.685428 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzrpf\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-kube-api-access-bzrpf\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.685573 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-registry-tls\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.685679 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-bound-sa-token\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.685801 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d82edf84-5eee-4dab-b5f7-459a986a8637-trusted-ca\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.686284 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d82edf84-5eee-4dab-b5f7-459a986a8637-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.687678 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d82edf84-5eee-4dab-b5f7-459a986a8637-trusted-ca\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.687716 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d82edf84-5eee-4dab-b5f7-459a986a8637-registry-certificates\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.695149 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d82edf84-5eee-4dab-b5f7-459a986a8637-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.696799 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-registry-tls\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.703682 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-bound-sa-token\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.704318 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzrpf\" (UniqueName: \"kubernetes.io/projected/d82edf84-5eee-4dab-b5f7-459a986a8637-kube-api-access-bzrpf\") pod \"image-registry-66df7c8f76-ckl57\" (UID: \"d82edf84-5eee-4dab-b5f7-459a986a8637\") " pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:46 crc kubenswrapper[4573]: I1203 08:43:46.809505 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:47 crc kubenswrapper[4573]: I1203 08:43:47.250463 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ckl57"] Dec 03 08:43:47 crc kubenswrapper[4573]: I1203 08:43:47.482392 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" event={"ID":"d82edf84-5eee-4dab-b5f7-459a986a8637","Type":"ContainerStarted","Data":"ec8287068474725c544d84fea26d523f935efaa0bd2b67604d1e23b1cbc665b8"} Dec 03 08:43:47 crc kubenswrapper[4573]: I1203 08:43:47.609281 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:47 crc kubenswrapper[4573]: I1203 08:43:47.646805 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 08:43:48 crc kubenswrapper[4573]: I1203 08:43:48.490735 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" event={"ID":"d82edf84-5eee-4dab-b5f7-459a986a8637","Type":"ContainerStarted","Data":"f2a824417ff497a58c84c4ea2179f0ef2eaec46d527da6589a209d5d34d0b150"} Dec 03 08:43:48 crc kubenswrapper[4573]: I1203 08:43:48.491276 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:43:48 crc kubenswrapper[4573]: I1203 08:43:48.512547 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" podStartSLOduration=2.5125266699999997 podStartE2EDuration="2.51252667s" podCreationTimestamp="2025-12-03 08:43:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:43:48.508769556 +0000 UTC m=+349.077148835" watchObservedRunningTime="2025-12-03 08:43:48.51252667 +0000 UTC m=+349.080905929" Dec 03 08:43:56 crc kubenswrapper[4573]: I1203 08:43:56.943641 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:43:56 crc kubenswrapper[4573]: I1203 08:43:56.943972 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:44:06 crc kubenswrapper[4573]: I1203 08:44:06.814543 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-ckl57" Dec 03 08:44:06 crc kubenswrapper[4573]: I1203 08:44:06.873220 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h7gd4"] Dec 03 08:44:26 crc kubenswrapper[4573]: I1203 08:44:26.943192 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:44:26 crc kubenswrapper[4573]: I1203 08:44:26.944131 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:44:31 crc kubenswrapper[4573]: I1203 08:44:31.920179 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" podUID="468b20da-e9e1-43e5-a265-54d131b1b66a" containerName="registry" containerID="cri-o://2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d" gracePeriod=30 Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.387468 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.585826 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"468b20da-e9e1-43e5-a265-54d131b1b66a\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.586180 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-bound-sa-token\") pod \"468b20da-e9e1-43e5-a265-54d131b1b66a\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.586223 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/468b20da-e9e1-43e5-a265-54d131b1b66a-installation-pull-secrets\") pod \"468b20da-e9e1-43e5-a265-54d131b1b66a\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.586303 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-tls\") pod \"468b20da-e9e1-43e5-a265-54d131b1b66a\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.587600 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6css7\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-kube-api-access-6css7\") pod \"468b20da-e9e1-43e5-a265-54d131b1b66a\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.587676 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/468b20da-e9e1-43e5-a265-54d131b1b66a-ca-trust-extracted\") pod \"468b20da-e9e1-43e5-a265-54d131b1b66a\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.587704 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-certificates\") pod \"468b20da-e9e1-43e5-a265-54d131b1b66a\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.587738 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-trusted-ca\") pod \"468b20da-e9e1-43e5-a265-54d131b1b66a\" (UID: \"468b20da-e9e1-43e5-a265-54d131b1b66a\") " Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.588611 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "468b20da-e9e1-43e5-a265-54d131b1b66a" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.589204 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "468b20da-e9e1-43e5-a265-54d131b1b66a" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.595025 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-kube-api-access-6css7" (OuterVolumeSpecName: "kube-api-access-6css7") pod "468b20da-e9e1-43e5-a265-54d131b1b66a" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a"). InnerVolumeSpecName "kube-api-access-6css7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.598583 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/468b20da-e9e1-43e5-a265-54d131b1b66a-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "468b20da-e9e1-43e5-a265-54d131b1b66a" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.603147 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "468b20da-e9e1-43e5-a265-54d131b1b66a" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.606871 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "468b20da-e9e1-43e5-a265-54d131b1b66a" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.607334 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "468b20da-e9e1-43e5-a265-54d131b1b66a" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.609228 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/468b20da-e9e1-43e5-a265-54d131b1b66a-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "468b20da-e9e1-43e5-a265-54d131b1b66a" (UID: "468b20da-e9e1-43e5-a265-54d131b1b66a"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.688569 4573 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/468b20da-e9e1-43e5-a265-54d131b1b66a-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.688613 4573 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.688625 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6css7\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-kube-api-access-6css7\") on node \"crc\" DevicePath \"\"" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.688635 4573 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/468b20da-e9e1-43e5-a265-54d131b1b66a-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.688648 4573 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.688657 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/468b20da-e9e1-43e5-a265-54d131b1b66a-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.688666 4573 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/468b20da-e9e1-43e5-a265-54d131b1b66a-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.799194 4573 generic.go:334] "Generic (PLEG): container finished" podID="468b20da-e9e1-43e5-a265-54d131b1b66a" containerID="2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d" exitCode=0 Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.799257 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" event={"ID":"468b20da-e9e1-43e5-a265-54d131b1b66a","Type":"ContainerDied","Data":"2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d"} Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.799301 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" event={"ID":"468b20da-e9e1-43e5-a265-54d131b1b66a","Type":"ContainerDied","Data":"47c2d81e0abf8d42085fe6c7147c17319d2e641aa99cb7405246348ee630bf39"} Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.799327 4573 scope.go:117] "RemoveContainer" containerID="2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.799490 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-h7gd4" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.831523 4573 scope.go:117] "RemoveContainer" containerID="2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d" Dec 03 08:44:32 crc kubenswrapper[4573]: E1203 08:44:32.832197 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d\": container with ID starting with 2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d not found: ID does not exist" containerID="2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.832276 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d"} err="failed to get container status \"2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d\": rpc error: code = NotFound desc = could not find container \"2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d\": container with ID starting with 2d2871bdbe424a09e764060fad0718899ede97438d10140737fad91f9e47a66d not found: ID does not exist" Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.833245 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h7gd4"] Dec 03 08:44:32 crc kubenswrapper[4573]: I1203 08:44:32.842469 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-h7gd4"] Dec 03 08:44:34 crc kubenswrapper[4573]: I1203 08:44:34.040328 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="468b20da-e9e1-43e5-a265-54d131b1b66a" path="/var/lib/kubelet/pods/468b20da-e9e1-43e5-a265-54d131b1b66a/volumes" Dec 03 08:44:56 crc kubenswrapper[4573]: I1203 08:44:56.944023 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:44:56 crc kubenswrapper[4573]: I1203 08:44:56.944967 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:44:56 crc kubenswrapper[4573]: I1203 08:44:56.945172 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:44:56 crc kubenswrapper[4573]: I1203 08:44:56.948147 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c69f0ae1788b1621dbf7e5b9317fb691e15bc289be6362f9912d1759f42dde21"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:44:56 crc kubenswrapper[4573]: I1203 08:44:56.948299 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://c69f0ae1788b1621dbf7e5b9317fb691e15bc289be6362f9912d1759f42dde21" gracePeriod=600 Dec 03 08:44:57 crc kubenswrapper[4573]: I1203 08:44:57.967163 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="c69f0ae1788b1621dbf7e5b9317fb691e15bc289be6362f9912d1759f42dde21" exitCode=0 Dec 03 08:44:57 crc kubenswrapper[4573]: I1203 08:44:57.967250 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"c69f0ae1788b1621dbf7e5b9317fb691e15bc289be6362f9912d1759f42dde21"} Dec 03 08:44:57 crc kubenswrapper[4573]: I1203 08:44:57.967830 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"0a70f7fefb2b62f5c62862c579bb68d0d1cc1089ab73a686b035e41c85444b01"} Dec 03 08:44:57 crc kubenswrapper[4573]: I1203 08:44:57.967866 4573 scope.go:117] "RemoveContainer" containerID="008c8e81cbafc803162d309625e34b587c5bf81a1eb4c3c9b48287d16234e99a" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.192027 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph"] Dec 03 08:45:00 crc kubenswrapper[4573]: E1203 08:45:00.193389 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="468b20da-e9e1-43e5-a265-54d131b1b66a" containerName="registry" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.193418 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="468b20da-e9e1-43e5-a265-54d131b1b66a" containerName="registry" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.193551 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="468b20da-e9e1-43e5-a265-54d131b1b66a" containerName="registry" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.194152 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.196606 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.196805 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.208398 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph"] Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.328275 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdkbr\" (UniqueName: \"kubernetes.io/projected/cc501733-940b-4735-8607-31f0bc1b1297-kube-api-access-rdkbr\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.329020 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc501733-940b-4735-8607-31f0bc1b1297-secret-volume\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.329109 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc501733-940b-4735-8607-31f0bc1b1297-config-volume\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.430037 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc501733-940b-4735-8607-31f0bc1b1297-secret-volume\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.430453 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc501733-940b-4735-8607-31f0bc1b1297-config-volume\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.430659 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdkbr\" (UniqueName: \"kubernetes.io/projected/cc501733-940b-4735-8607-31f0bc1b1297-kube-api-access-rdkbr\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.431745 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc501733-940b-4735-8607-31f0bc1b1297-config-volume\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.442587 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc501733-940b-4735-8607-31f0bc1b1297-secret-volume\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.453799 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdkbr\" (UniqueName: \"kubernetes.io/projected/cc501733-940b-4735-8607-31f0bc1b1297-kube-api-access-rdkbr\") pod \"collect-profiles-29412525-rvqph\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.521391 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.777097 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph"] Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.991076 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" event={"ID":"cc501733-940b-4735-8607-31f0bc1b1297","Type":"ContainerStarted","Data":"3be2fafc5e13740d824be85dada8f6f0568fbfa608c67f34f15419d1a5bc037a"} Dec 03 08:45:00 crc kubenswrapper[4573]: I1203 08:45:00.991132 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" event={"ID":"cc501733-940b-4735-8607-31f0bc1b1297","Type":"ContainerStarted","Data":"c46b69620293d24c8fe491660b52e9b5a8a81b3038abcacd114dd2d9ccc7b514"} Dec 03 08:45:01 crc kubenswrapper[4573]: I1203 08:45:01.010869 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" podStartSLOduration=1.010835834 podStartE2EDuration="1.010835834s" podCreationTimestamp="2025-12-03 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:45:01.005818795 +0000 UTC m=+421.574198064" watchObservedRunningTime="2025-12-03 08:45:01.010835834 +0000 UTC m=+421.579215093" Dec 03 08:45:02 crc kubenswrapper[4573]: I1203 08:45:02.001131 4573 generic.go:334] "Generic (PLEG): container finished" podID="cc501733-940b-4735-8607-31f0bc1b1297" containerID="3be2fafc5e13740d824be85dada8f6f0568fbfa608c67f34f15419d1a5bc037a" exitCode=0 Dec 03 08:45:02 crc kubenswrapper[4573]: I1203 08:45:02.001225 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" event={"ID":"cc501733-940b-4735-8607-31f0bc1b1297","Type":"ContainerDied","Data":"3be2fafc5e13740d824be85dada8f6f0568fbfa608c67f34f15419d1a5bc037a"} Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.269796 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.285260 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc501733-940b-4735-8607-31f0bc1b1297-secret-volume\") pod \"cc501733-940b-4735-8607-31f0bc1b1297\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.285380 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc501733-940b-4735-8607-31f0bc1b1297-config-volume\") pod \"cc501733-940b-4735-8607-31f0bc1b1297\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.285434 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdkbr\" (UniqueName: \"kubernetes.io/projected/cc501733-940b-4735-8607-31f0bc1b1297-kube-api-access-rdkbr\") pod \"cc501733-940b-4735-8607-31f0bc1b1297\" (UID: \"cc501733-940b-4735-8607-31f0bc1b1297\") " Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.287298 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc501733-940b-4735-8607-31f0bc1b1297-config-volume" (OuterVolumeSpecName: "config-volume") pod "cc501733-940b-4735-8607-31f0bc1b1297" (UID: "cc501733-940b-4735-8607-31f0bc1b1297"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.292962 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc501733-940b-4735-8607-31f0bc1b1297-kube-api-access-rdkbr" (OuterVolumeSpecName: "kube-api-access-rdkbr") pod "cc501733-940b-4735-8607-31f0bc1b1297" (UID: "cc501733-940b-4735-8607-31f0bc1b1297"). InnerVolumeSpecName "kube-api-access-rdkbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.297030 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc501733-940b-4735-8607-31f0bc1b1297-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cc501733-940b-4735-8607-31f0bc1b1297" (UID: "cc501733-940b-4735-8607-31f0bc1b1297"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.386947 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc501733-940b-4735-8607-31f0bc1b1297-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.386990 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdkbr\" (UniqueName: \"kubernetes.io/projected/cc501733-940b-4735-8607-31f0bc1b1297-kube-api-access-rdkbr\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:03 crc kubenswrapper[4573]: I1203 08:45:03.387003 4573 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc501733-940b-4735-8607-31f0bc1b1297-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 08:45:04 crc kubenswrapper[4573]: I1203 08:45:04.019932 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" event={"ID":"cc501733-940b-4735-8607-31f0bc1b1297","Type":"ContainerDied","Data":"c46b69620293d24c8fe491660b52e9b5a8a81b3038abcacd114dd2d9ccc7b514"} Dec 03 08:45:04 crc kubenswrapper[4573]: I1203 08:45:04.020417 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c46b69620293d24c8fe491660b52e9b5a8a81b3038abcacd114dd2d9ccc7b514" Dec 03 08:45:04 crc kubenswrapper[4573]: I1203 08:45:04.020033 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph" Dec 03 08:47:26 crc kubenswrapper[4573]: I1203 08:47:26.943101 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:47:26 crc kubenswrapper[4573]: I1203 08:47:26.944349 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:47:56 crc kubenswrapper[4573]: I1203 08:47:56.943698 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:47:56 crc kubenswrapper[4573]: I1203 08:47:56.945213 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:48:26 crc kubenswrapper[4573]: I1203 08:48:26.943325 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:48:26 crc kubenswrapper[4573]: I1203 08:48:26.944285 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:48:26 crc kubenswrapper[4573]: I1203 08:48:26.944349 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:48:26 crc kubenswrapper[4573]: I1203 08:48:26.945145 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a70f7fefb2b62f5c62862c579bb68d0d1cc1089ab73a686b035e41c85444b01"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:48:26 crc kubenswrapper[4573]: I1203 08:48:26.945213 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://0a70f7fefb2b62f5c62862c579bb68d0d1cc1089ab73a686b035e41c85444b01" gracePeriod=600 Dec 03 08:48:27 crc kubenswrapper[4573]: I1203 08:48:27.361511 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="0a70f7fefb2b62f5c62862c579bb68d0d1cc1089ab73a686b035e41c85444b01" exitCode=0 Dec 03 08:48:27 crc kubenswrapper[4573]: I1203 08:48:27.361561 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"0a70f7fefb2b62f5c62862c579bb68d0d1cc1089ab73a686b035e41c85444b01"} Dec 03 08:48:27 crc kubenswrapper[4573]: I1203 08:48:27.362033 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"e638f4437cd42f13a1acdd69aa087a2d6a308b827f8a1fbde0eba4442f1d564e"} Dec 03 08:48:27 crc kubenswrapper[4573]: I1203 08:48:27.362072 4573 scope.go:117] "RemoveContainer" containerID="c69f0ae1788b1621dbf7e5b9317fb691e15bc289be6362f9912d1759f42dde21" Dec 03 08:50:40 crc kubenswrapper[4573]: I1203 08:50:40.898172 4573 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 08:50:56 crc kubenswrapper[4573]: I1203 08:50:56.943243 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:50:56 crc kubenswrapper[4573]: I1203 08:50:56.945235 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:51:26 crc kubenswrapper[4573]: I1203 08:51:26.943747 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:51:26 crc kubenswrapper[4573]: I1203 08:51:26.944712 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:51:56 crc kubenswrapper[4573]: I1203 08:51:56.943033 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:51:56 crc kubenswrapper[4573]: I1203 08:51:56.943940 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:51:56 crc kubenswrapper[4573]: I1203 08:51:56.944007 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:51:56 crc kubenswrapper[4573]: I1203 08:51:56.944765 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e638f4437cd42f13a1acdd69aa087a2d6a308b827f8a1fbde0eba4442f1d564e"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:51:56 crc kubenswrapper[4573]: I1203 08:51:56.944827 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://e638f4437cd42f13a1acdd69aa087a2d6a308b827f8a1fbde0eba4442f1d564e" gracePeriod=600 Dec 03 08:51:57 crc kubenswrapper[4573]: I1203 08:51:57.749229 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="e638f4437cd42f13a1acdd69aa087a2d6a308b827f8a1fbde0eba4442f1d564e" exitCode=0 Dec 03 08:51:57 crc kubenswrapper[4573]: I1203 08:51:57.749314 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"e638f4437cd42f13a1acdd69aa087a2d6a308b827f8a1fbde0eba4442f1d564e"} Dec 03 08:51:57 crc kubenswrapper[4573]: I1203 08:51:57.749972 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"68d9133084811f469a486ce4e9540b723fb96d8cd528443eac595b0a1c6329ed"} Dec 03 08:51:57 crc kubenswrapper[4573]: I1203 08:51:57.749998 4573 scope.go:117] "RemoveContainer" containerID="0a70f7fefb2b62f5c62862c579bb68d0d1cc1089ab73a686b035e41c85444b01" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.626185 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-h9bp4"] Dec 03 08:52:18 crc kubenswrapper[4573]: E1203 08:52:18.627515 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc501733-940b-4735-8607-31f0bc1b1297" containerName="collect-profiles" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.627541 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc501733-940b-4735-8607-31f0bc1b1297" containerName="collect-profiles" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.627708 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc501733-940b-4735-8607-31f0bc1b1297" containerName="collect-profiles" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.628463 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-h9bp4" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.630557 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4226f"] Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.631479 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4226f" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.636611 4573 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-mbh5v" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.646712 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.650580 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.651212 4573 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-6mp75" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.691698 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-twvqd"] Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.699357 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.735444 4573 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-b4ftp" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.735775 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-h9bp4"] Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.767185 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4226f"] Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.794773 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-twvqd"] Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.819597 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6xl8\" (UniqueName: \"kubernetes.io/projected/f4edafce-1ec1-4ada-9861-c36ce8c3906f-kube-api-access-g6xl8\") pod \"cert-manager-webhook-5655c58dd6-twvqd\" (UID: \"f4edafce-1ec1-4ada-9861-c36ce8c3906f\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.819686 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64fxk\" (UniqueName: \"kubernetes.io/projected/9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf-kube-api-access-64fxk\") pod \"cert-manager-cainjector-7f985d654d-4226f\" (UID: \"9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4226f" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.819712 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7qqg\" (UniqueName: \"kubernetes.io/projected/ee0f7af7-48fb-433a-8f6b-801fd53f331d-kube-api-access-b7qqg\") pod \"cert-manager-5b446d88c5-h9bp4\" (UID: \"ee0f7af7-48fb-433a-8f6b-801fd53f331d\") " pod="cert-manager/cert-manager-5b446d88c5-h9bp4" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.921015 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6xl8\" (UniqueName: \"kubernetes.io/projected/f4edafce-1ec1-4ada-9861-c36ce8c3906f-kube-api-access-g6xl8\") pod \"cert-manager-webhook-5655c58dd6-twvqd\" (UID: \"f4edafce-1ec1-4ada-9861-c36ce8c3906f\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.921595 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64fxk\" (UniqueName: \"kubernetes.io/projected/9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf-kube-api-access-64fxk\") pod \"cert-manager-cainjector-7f985d654d-4226f\" (UID: \"9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4226f" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.921627 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7qqg\" (UniqueName: \"kubernetes.io/projected/ee0f7af7-48fb-433a-8f6b-801fd53f331d-kube-api-access-b7qqg\") pod \"cert-manager-5b446d88c5-h9bp4\" (UID: \"ee0f7af7-48fb-433a-8f6b-801fd53f331d\") " pod="cert-manager/cert-manager-5b446d88c5-h9bp4" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.944722 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7qqg\" (UniqueName: \"kubernetes.io/projected/ee0f7af7-48fb-433a-8f6b-801fd53f331d-kube-api-access-b7qqg\") pod \"cert-manager-5b446d88c5-h9bp4\" (UID: \"ee0f7af7-48fb-433a-8f6b-801fd53f331d\") " pod="cert-manager/cert-manager-5b446d88c5-h9bp4" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.944815 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64fxk\" (UniqueName: \"kubernetes.io/projected/9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf-kube-api-access-64fxk\") pod \"cert-manager-cainjector-7f985d654d-4226f\" (UID: \"9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4226f" Dec 03 08:52:18 crc kubenswrapper[4573]: I1203 08:52:18.945161 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6xl8\" (UniqueName: \"kubernetes.io/projected/f4edafce-1ec1-4ada-9861-c36ce8c3906f-kube-api-access-g6xl8\") pod \"cert-manager-webhook-5655c58dd6-twvqd\" (UID: \"f4edafce-1ec1-4ada-9861-c36ce8c3906f\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.036851 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-h9bp4" Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.062577 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4226f" Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.082170 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.383927 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-h9bp4"] Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.393174 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.428182 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4226f"] Dec 03 08:52:19 crc kubenswrapper[4573]: W1203 08:52:19.434435 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d7c2bf5_9ddc_4b6e_9aa3_f5cef849aedf.slice/crio-baa213c11e73711448a81595a92694a2c7014cd97422975ce1a985f538c71efa WatchSource:0}: Error finding container baa213c11e73711448a81595a92694a2c7014cd97422975ce1a985f538c71efa: Status 404 returned error can't find the container with id baa213c11e73711448a81595a92694a2c7014cd97422975ce1a985f538c71efa Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.477891 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-twvqd"] Dec 03 08:52:19 crc kubenswrapper[4573]: W1203 08:52:19.486204 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4edafce_1ec1_4ada_9861_c36ce8c3906f.slice/crio-b0cad3b8bb2c281c1ce6db5db4e12a4e394dd6e57dfd5c9868c430b72a0dc3c3 WatchSource:0}: Error finding container b0cad3b8bb2c281c1ce6db5db4e12a4e394dd6e57dfd5c9868c430b72a0dc3c3: Status 404 returned error can't find the container with id b0cad3b8bb2c281c1ce6db5db4e12a4e394dd6e57dfd5c9868c430b72a0dc3c3 Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.905521 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-h9bp4" event={"ID":"ee0f7af7-48fb-433a-8f6b-801fd53f331d","Type":"ContainerStarted","Data":"81e2770746e7b9469f196ae23149732dbc6442c3b40cb1b3bd44e82fa4c9fcf4"} Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.906552 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" event={"ID":"f4edafce-1ec1-4ada-9861-c36ce8c3906f","Type":"ContainerStarted","Data":"b0cad3b8bb2c281c1ce6db5db4e12a4e394dd6e57dfd5c9868c430b72a0dc3c3"} Dec 03 08:52:19 crc kubenswrapper[4573]: I1203 08:52:19.908514 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4226f" event={"ID":"9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf","Type":"ContainerStarted","Data":"baa213c11e73711448a81595a92694a2c7014cd97422975ce1a985f538c71efa"} Dec 03 08:52:23 crc kubenswrapper[4573]: I1203 08:52:23.936609 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" event={"ID":"f4edafce-1ec1-4ada-9861-c36ce8c3906f","Type":"ContainerStarted","Data":"737b6d0dfd76a1371ecb52cbef02e5fc5b2ee3e64f814c9d05712fb7bd5f76c1"} Dec 03 08:52:23 crc kubenswrapper[4573]: I1203 08:52:23.937456 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" Dec 03 08:52:23 crc kubenswrapper[4573]: I1203 08:52:23.940534 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4226f" event={"ID":"9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf","Type":"ContainerStarted","Data":"1f471f194d53bd6ff3eda7d3ff68bd1145324f2e964b186ef884f5f2899cc580"} Dec 03 08:52:23 crc kubenswrapper[4573]: I1203 08:52:23.942153 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-h9bp4" event={"ID":"ee0f7af7-48fb-433a-8f6b-801fd53f331d","Type":"ContainerStarted","Data":"6f9f76a99f97c82ceb5cab68b7e0167ce4f847b433c35eca40da9f4a9e8da89e"} Dec 03 08:52:23 crc kubenswrapper[4573]: I1203 08:52:23.985878 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" podStartSLOduration=2.055232443 podStartE2EDuration="5.985845326s" podCreationTimestamp="2025-12-03 08:52:18 +0000 UTC" firstStartedPulling="2025-12-03 08:52:19.494318135 +0000 UTC m=+860.062697394" lastFinishedPulling="2025-12-03 08:52:23.424931018 +0000 UTC m=+863.993310277" observedRunningTime="2025-12-03 08:52:23.963258351 +0000 UTC m=+864.531637610" watchObservedRunningTime="2025-12-03 08:52:23.985845326 +0000 UTC m=+864.554224585" Dec 03 08:52:23 crc kubenswrapper[4573]: I1203 08:52:23.986882 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-h9bp4" podStartSLOduration=2.037330966 podStartE2EDuration="5.986873663s" podCreationTimestamp="2025-12-03 08:52:18 +0000 UTC" firstStartedPulling="2025-12-03 08:52:19.392781032 +0000 UTC m=+859.961160291" lastFinishedPulling="2025-12-03 08:52:23.342323729 +0000 UTC m=+863.910702988" observedRunningTime="2025-12-03 08:52:23.980816849 +0000 UTC m=+864.549196108" watchObservedRunningTime="2025-12-03 08:52:23.986873663 +0000 UTC m=+864.555252922" Dec 03 08:52:24 crc kubenswrapper[4573]: I1203 08:52:24.012590 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-4226f" podStartSLOduration=2.117086358 podStartE2EDuration="6.012553302s" podCreationTimestamp="2025-12-03 08:52:18 +0000 UTC" firstStartedPulling="2025-12-03 08:52:19.439093683 +0000 UTC m=+860.007472942" lastFinishedPulling="2025-12-03 08:52:23.334560627 +0000 UTC m=+863.902939886" observedRunningTime="2025-12-03 08:52:24.009437587 +0000 UTC m=+864.577816856" watchObservedRunningTime="2025-12-03 08:52:24.012553302 +0000 UTC m=+864.580932581" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.389968 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x466z"] Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.391490 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovn-controller" containerID="cri-o://8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8" gracePeriod=30 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.391524 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="nbdb" containerID="cri-o://c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e" gracePeriod=30 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.391674 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="sbdb" containerID="cri-o://b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5" gracePeriod=30 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.391736 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c" gracePeriod=30 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.391782 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="northd" containerID="cri-o://517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7" gracePeriod=30 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.391822 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kube-rbac-proxy-node" containerID="cri-o://3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216" gracePeriod=30 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.391860 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovn-acl-logging" containerID="cri-o://7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73" gracePeriod=30 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.457129 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" containerID="cri-o://ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599" gracePeriod=30 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.769612 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/3.log" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.772344 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovn-acl-logging/0.log" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.772911 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovn-controller/0.log" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.773370 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.900982 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-systemd\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901060 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-systemd-units\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901102 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48186949-d112-4fb7-aa60-026d0176c73e-ovn-node-metrics-cert\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901128 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-netd\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901177 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-config\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901181 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901204 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-slash\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901254 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-var-lib-openvswitch\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901271 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901283 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-env-overrides\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901330 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-script-lib\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901353 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-var-lib-cni-networks-ovn-kubernetes\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901389 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-bin\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901415 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-openvswitch\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901431 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-etc-openvswitch\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901474 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-ovn\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901504 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-log-socket\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901542 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fssxg\" (UniqueName: \"kubernetes.io/projected/48186949-d112-4fb7-aa60-026d0176c73e-kube-api-access-fssxg\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901566 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-kubelet\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901652 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-ovn-kubernetes\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901692 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-node-log\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.901711 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-netns\") pod \"48186949-d112-4fb7-aa60-026d0176c73e\" (UID: \"48186949-d112-4fb7-aa60-026d0176c73e\") " Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.902117 4573 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.902129 4573 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.902166 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.902636 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.902780 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-slash" (OuterVolumeSpecName: "host-slash") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.902816 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903175 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903390 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903681 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903724 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903782 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903814 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-log-socket" (OuterVolumeSpecName: "log-socket") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903816 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903838 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903862 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.903888 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-node-log" (OuterVolumeSpecName: "node-log") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.907137 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.924925 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48186949-d112-4fb7-aa60-026d0176c73e-kube-api-access-fssxg" (OuterVolumeSpecName: "kube-api-access-fssxg") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "kube-api-access-fssxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.938241 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48186949-d112-4fb7-aa60-026d0176c73e-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.942697 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "48186949-d112-4fb7-aa60-026d0176c73e" (UID: "48186949-d112-4fb7-aa60-026d0176c73e"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964425 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hb68s"] Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964659 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kube-rbac-proxy-node" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964673 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kube-rbac-proxy-node" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964684 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964690 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964697 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964703 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964711 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovn-acl-logging" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964717 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovn-acl-logging" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964725 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964731 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964739 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kubecfg-setup" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964745 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kubecfg-setup" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964752 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="nbdb" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964757 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="nbdb" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964765 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964771 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964779 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovn-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964786 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovn-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964799 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964806 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964814 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="sbdb" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964821 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="sbdb" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.964832 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="northd" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964839 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="northd" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964938 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="sbdb" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964949 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="northd" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964957 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kube-rbac-proxy-node" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964967 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964973 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964980 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="nbdb" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964989 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.964996 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.965004 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.965012 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovn-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.965019 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovn-acl-logging" Dec 03 08:52:28 crc kubenswrapper[4573]: E1203 08:52:28.965235 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.965245 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.965343 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="48186949-d112-4fb7-aa60-026d0176c73e" containerName="ovnkube-controller" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.967098 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.980186 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/2.log" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.980772 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/1.log" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.980838 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c98ef96-0e91-49e3-88d5-c95dd6f52c5d" containerID="dcf9f6f434fdad7cbbf3d2a5bda9998bad92078f8bb45c93cdb1b3fc856089c1" exitCode=2 Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.980933 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ch24s" event={"ID":"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d","Type":"ContainerDied","Data":"dcf9f6f434fdad7cbbf3d2a5bda9998bad92078f8bb45c93cdb1b3fc856089c1"} Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.980995 4573 scope.go:117] "RemoveContainer" containerID="35a4fe7bc5dfd0ccc3be381e02a0e55aec5954d8303f03ef60e537d98eb38ab0" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.982156 4573 scope.go:117] "RemoveContainer" containerID="dcf9f6f434fdad7cbbf3d2a5bda9998bad92078f8bb45c93cdb1b3fc856089c1" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.992425 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovnkube-controller/3.log" Dec 03 08:52:28 crc kubenswrapper[4573]: I1203 08:52:28.999493 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovn-acl-logging/0.log" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:28.999994 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x466z_48186949-d112-4fb7-aa60-026d0176c73e/ovn-controller/0.log" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000376 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599" exitCode=0 Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000414 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5" exitCode=0 Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000426 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e" exitCode=0 Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000435 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7" exitCode=0 Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000445 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c" exitCode=0 Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000455 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216" exitCode=0 Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000464 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73" exitCode=143 Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000473 4573 generic.go:334] "Generic (PLEG): container finished" podID="48186949-d112-4fb7-aa60-026d0176c73e" containerID="8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8" exitCode=143 Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000506 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000553 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000571 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000588 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000602 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000615 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000632 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000649 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000659 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000667 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000673 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000682 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000690 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000698 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000705 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000713 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000724 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000736 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000745 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000754 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000761 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000769 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000777 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000785 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000792 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000798 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000805 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000816 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000828 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000837 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000844 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000851 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000859 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000867 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000875 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000882 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000889 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000896 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000906 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" event={"ID":"48186949-d112-4fb7-aa60-026d0176c73e","Type":"ContainerDied","Data":"840a3cbc3222906512a6e6cfd20c7c4c9e8d20852872eaeab26156a645e04825"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000918 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000930 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000938 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000945 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000952 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000960 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000969 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000976 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000983 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.000991 4573 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.001294 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x466z" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.004932 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovnkube-script-lib\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.004984 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-ovn\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005019 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-cni-bin\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005038 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovn-node-metrics-cert\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005074 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-etc-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005120 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-log-socket\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005159 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-run-netns\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005191 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-systemd-units\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005208 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4hdl\" (UniqueName: \"kubernetes.io/projected/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-kube-api-access-c4hdl\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005223 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005238 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-run-ovn-kubernetes\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005254 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-slash\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005292 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovnkube-config\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005312 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-cni-netd\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005342 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-systemd\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005371 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-node-log\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005393 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005436 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-var-lib-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005456 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-kubelet\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005480 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-env-overrides\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005539 4573 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005554 4573 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005567 4573 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005579 4573 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005591 4573 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48186949-d112-4fb7-aa60-026d0176c73e-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005601 4573 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005610 4573 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005620 4573 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005629 4573 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005639 4573 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48186949-d112-4fb7-aa60-026d0176c73e-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005649 4573 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005659 4573 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005669 4573 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005678 4573 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005688 4573 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005697 4573 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005706 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fssxg\" (UniqueName: \"kubernetes.io/projected/48186949-d112-4fb7-aa60-026d0176c73e-kube-api-access-fssxg\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.005716 4573 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48186949-d112-4fb7-aa60-026d0176c73e-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.036653 4573 scope.go:117] "RemoveContainer" containerID="ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.085245 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.092461 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-twvqd" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.107992 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-cni-bin\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108066 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovn-node-metrics-cert\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108110 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-etc-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108134 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-cni-bin\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108144 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-log-socket\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108240 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-etc-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108251 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-run-netns\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108290 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-run-netns\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108303 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-systemd-units\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108326 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4hdl\" (UniqueName: \"kubernetes.io/projected/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-kube-api-access-c4hdl\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108358 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108373 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-run-ovn-kubernetes\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108401 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-slash\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108427 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovnkube-config\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108447 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-cni-netd\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108467 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-systemd\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108486 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-node-log\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108506 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108540 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-var-lib-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108558 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-kubelet\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108577 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-env-overrides\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108600 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovnkube-script-lib\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108617 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-ovn\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108674 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-ovn\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108698 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-cni-netd\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108712 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-systemd\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108750 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-node-log\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108755 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-systemd-units\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108778 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108809 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-var-lib-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.108829 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-kubelet\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.109421 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-env-overrides\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.109568 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-run-openvswitch\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.109583 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-slash\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.109619 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-host-run-ovn-kubernetes\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.109935 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovnkube-script-lib\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.109994 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-log-socket\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.110689 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovnkube-config\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.111964 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-ovn-node-metrics-cert\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.118233 4573 scope.go:117] "RemoveContainer" containerID="b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.123221 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x466z"] Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.132677 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4hdl\" (UniqueName: \"kubernetes.io/projected/200c1dec-7901-4fe2-91d2-e4ad1ecbedf7-kube-api-access-c4hdl\") pod \"ovnkube-node-hb68s\" (UID: \"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.138508 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x466z"] Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.153759 4573 scope.go:117] "RemoveContainer" containerID="c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.174437 4573 scope.go:117] "RemoveContainer" containerID="517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.198971 4573 scope.go:117] "RemoveContainer" containerID="c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.215380 4573 scope.go:117] "RemoveContainer" containerID="3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.231940 4573 scope.go:117] "RemoveContainer" containerID="7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.253127 4573 scope.go:117] "RemoveContainer" containerID="8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.273107 4573 scope.go:117] "RemoveContainer" containerID="727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.288983 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.301183 4573 scope.go:117] "RemoveContainer" containerID="ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.302076 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": container with ID starting with ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599 not found: ID does not exist" containerID="ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.302147 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} err="failed to get container status \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": rpc error: code = NotFound desc = could not find container \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": container with ID starting with ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.302185 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.302638 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": container with ID starting with 01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1 not found: ID does not exist" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.302675 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} err="failed to get container status \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": rpc error: code = NotFound desc = could not find container \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": container with ID starting with 01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.302714 4573 scope.go:117] "RemoveContainer" containerID="b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.303326 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": container with ID starting with b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5 not found: ID does not exist" containerID="b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.303355 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} err="failed to get container status \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": rpc error: code = NotFound desc = could not find container \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": container with ID starting with b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.303371 4573 scope.go:117] "RemoveContainer" containerID="c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.304013 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": container with ID starting with c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e not found: ID does not exist" containerID="c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.304974 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} err="failed to get container status \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": rpc error: code = NotFound desc = could not find container \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": container with ID starting with c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.305034 4573 scope.go:117] "RemoveContainer" containerID="517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.305399 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": container with ID starting with 517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7 not found: ID does not exist" containerID="517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.305426 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} err="failed to get container status \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": rpc error: code = NotFound desc = could not find container \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": container with ID starting with 517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.305441 4573 scope.go:117] "RemoveContainer" containerID="c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.305689 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": container with ID starting with c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c not found: ID does not exist" containerID="c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.305727 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} err="failed to get container status \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": rpc error: code = NotFound desc = could not find container \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": container with ID starting with c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.305745 4573 scope.go:117] "RemoveContainer" containerID="3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.305976 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": container with ID starting with 3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216 not found: ID does not exist" containerID="3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.306004 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} err="failed to get container status \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": rpc error: code = NotFound desc = could not find container \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": container with ID starting with 3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.306017 4573 scope.go:117] "RemoveContainer" containerID="7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.306259 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": container with ID starting with 7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73 not found: ID does not exist" containerID="7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.306279 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} err="failed to get container status \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": rpc error: code = NotFound desc = could not find container \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": container with ID starting with 7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.306294 4573 scope.go:117] "RemoveContainer" containerID="8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.306642 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": container with ID starting with 8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8 not found: ID does not exist" containerID="8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.306692 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} err="failed to get container status \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": rpc error: code = NotFound desc = could not find container \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": container with ID starting with 8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.306715 4573 scope.go:117] "RemoveContainer" containerID="727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade" Dec 03 08:52:29 crc kubenswrapper[4573]: E1203 08:52:29.310254 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": container with ID starting with 727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade not found: ID does not exist" containerID="727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.310287 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} err="failed to get container status \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": rpc error: code = NotFound desc = could not find container \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": container with ID starting with 727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.310308 4573 scope.go:117] "RemoveContainer" containerID="ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.310555 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} err="failed to get container status \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": rpc error: code = NotFound desc = could not find container \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": container with ID starting with ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.310575 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.310777 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} err="failed to get container status \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": rpc error: code = NotFound desc = could not find container \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": container with ID starting with 01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.310795 4573 scope.go:117] "RemoveContainer" containerID="b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311011 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} err="failed to get container status \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": rpc error: code = NotFound desc = could not find container \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": container with ID starting with b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311024 4573 scope.go:117] "RemoveContainer" containerID="c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311202 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} err="failed to get container status \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": rpc error: code = NotFound desc = could not find container \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": container with ID starting with c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311215 4573 scope.go:117] "RemoveContainer" containerID="517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311366 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} err="failed to get container status \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": rpc error: code = NotFound desc = could not find container \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": container with ID starting with 517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311381 4573 scope.go:117] "RemoveContainer" containerID="c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311546 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} err="failed to get container status \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": rpc error: code = NotFound desc = could not find container \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": container with ID starting with c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311563 4573 scope.go:117] "RemoveContainer" containerID="3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311734 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} err="failed to get container status \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": rpc error: code = NotFound desc = could not find container \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": container with ID starting with 3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311749 4573 scope.go:117] "RemoveContainer" containerID="7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311906 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} err="failed to get container status \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": rpc error: code = NotFound desc = could not find container \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": container with ID starting with 7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.311922 4573 scope.go:117] "RemoveContainer" containerID="8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.312094 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} err="failed to get container status \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": rpc error: code = NotFound desc = could not find container \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": container with ID starting with 8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.312109 4573 scope.go:117] "RemoveContainer" containerID="727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.312278 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} err="failed to get container status \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": rpc error: code = NotFound desc = could not find container \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": container with ID starting with 727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.312303 4573 scope.go:117] "RemoveContainer" containerID="ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.312465 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} err="failed to get container status \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": rpc error: code = NotFound desc = could not find container \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": container with ID starting with ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.312481 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.312634 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} err="failed to get container status \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": rpc error: code = NotFound desc = could not find container \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": container with ID starting with 01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.312652 4573 scope.go:117] "RemoveContainer" containerID="b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.313883 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} err="failed to get container status \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": rpc error: code = NotFound desc = could not find container \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": container with ID starting with b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.313980 4573 scope.go:117] "RemoveContainer" containerID="c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.314526 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} err="failed to get container status \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": rpc error: code = NotFound desc = could not find container \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": container with ID starting with c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.314547 4573 scope.go:117] "RemoveContainer" containerID="517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.314742 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} err="failed to get container status \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": rpc error: code = NotFound desc = could not find container \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": container with ID starting with 517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.314763 4573 scope.go:117] "RemoveContainer" containerID="c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.314957 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} err="failed to get container status \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": rpc error: code = NotFound desc = could not find container \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": container with ID starting with c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.314980 4573 scope.go:117] "RemoveContainer" containerID="3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.315340 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} err="failed to get container status \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": rpc error: code = NotFound desc = could not find container \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": container with ID starting with 3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.315392 4573 scope.go:117] "RemoveContainer" containerID="7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.316043 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} err="failed to get container status \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": rpc error: code = NotFound desc = could not find container \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": container with ID starting with 7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.316145 4573 scope.go:117] "RemoveContainer" containerID="8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.316607 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} err="failed to get container status \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": rpc error: code = NotFound desc = could not find container \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": container with ID starting with 8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.316632 4573 scope.go:117] "RemoveContainer" containerID="727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.317105 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} err="failed to get container status \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": rpc error: code = NotFound desc = could not find container \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": container with ID starting with 727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.317227 4573 scope.go:117] "RemoveContainer" containerID="ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.318326 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599"} err="failed to get container status \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": rpc error: code = NotFound desc = could not find container \"ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599\": container with ID starting with ff9295f4792058b7d7364cd651f561abe9ea7cbdac4e98c805ef4b253596b599 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.318366 4573 scope.go:117] "RemoveContainer" containerID="01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.318810 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1"} err="failed to get container status \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": rpc error: code = NotFound desc = could not find container \"01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1\": container with ID starting with 01573b5ba71b64f8230f1cc9536eae8a7293fe1a062c37a15afe8dbe3e3d43d1 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.318844 4573 scope.go:117] "RemoveContainer" containerID="b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.319316 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5"} err="failed to get container status \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": rpc error: code = NotFound desc = could not find container \"b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5\": container with ID starting with b1fbfa55e248e3026042bc44d9382121fe4866e2bd32ed4c2bf08445eb4c05a5 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.319473 4573 scope.go:117] "RemoveContainer" containerID="c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.320915 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e"} err="failed to get container status \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": rpc error: code = NotFound desc = could not find container \"c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e\": container with ID starting with c6b64634da65a9b343f8cb023359a4b9f33d24a9a026ad9e45c32b9abea94b1e not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.320942 4573 scope.go:117] "RemoveContainer" containerID="517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.321214 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7"} err="failed to get container status \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": rpc error: code = NotFound desc = could not find container \"517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7\": container with ID starting with 517277d8d1794c71e8a520131c23e06a1afe78750d45341adfd0936273eb2dd7 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.321233 4573 scope.go:117] "RemoveContainer" containerID="c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.321478 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c"} err="failed to get container status \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": rpc error: code = NotFound desc = could not find container \"c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c\": container with ID starting with c490bf938d7f9cbeec2e953cd109ed5fcaf818da51b62faea9ccd5c4b9bb766c not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.321501 4573 scope.go:117] "RemoveContainer" containerID="3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.321788 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216"} err="failed to get container status \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": rpc error: code = NotFound desc = could not find container \"3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216\": container with ID starting with 3fb05de48ee180d9831aa46b1cee3fdc3e38516e815b48e0c1b6906e15fb3216 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.321815 4573 scope.go:117] "RemoveContainer" containerID="7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.322327 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73"} err="failed to get container status \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": rpc error: code = NotFound desc = could not find container \"7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73\": container with ID starting with 7d853838c9e1e256e0a200f6ae8bd6f98748c66bd98defa4eb7933ca15f0ab73 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.322414 4573 scope.go:117] "RemoveContainer" containerID="8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.323153 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8"} err="failed to get container status \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": rpc error: code = NotFound desc = could not find container \"8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8\": container with ID starting with 8f08f0eaa8c07a51841e881ad183e845a06ff86e0382c92f754f776600d1a1c8 not found: ID does not exist" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.323184 4573 scope.go:117] "RemoveContainer" containerID="727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade" Dec 03 08:52:29 crc kubenswrapper[4573]: I1203 08:52:29.323528 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade"} err="failed to get container status \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": rpc error: code = NotFound desc = could not find container \"727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade\": container with ID starting with 727329a67165ee144ebe8813f264a30b0f7c4d482f2e4b1c3a3482a6f52f8ade not found: ID does not exist" Dec 03 08:52:30 crc kubenswrapper[4573]: I1203 08:52:30.013485 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-ch24s_8c98ef96-0e91-49e3-88d5-c95dd6f52c5d/kube-multus/2.log" Dec 03 08:52:30 crc kubenswrapper[4573]: I1203 08:52:30.014287 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-ch24s" event={"ID":"8c98ef96-0e91-49e3-88d5-c95dd6f52c5d","Type":"ContainerStarted","Data":"2fe1f70b640dfc47551b6346a3421553b1a7946a1acec6059cee13ef802761f6"} Dec 03 08:52:30 crc kubenswrapper[4573]: I1203 08:52:30.021317 4573 generic.go:334] "Generic (PLEG): container finished" podID="200c1dec-7901-4fe2-91d2-e4ad1ecbedf7" containerID="b6aca7b183525cf72100f212f6c4efff80e2bc66d54933b79d8cd4c6d35f7d3e" exitCode=0 Dec 03 08:52:30 crc kubenswrapper[4573]: I1203 08:52:30.021403 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerDied","Data":"b6aca7b183525cf72100f212f6c4efff80e2bc66d54933b79d8cd4c6d35f7d3e"} Dec 03 08:52:30 crc kubenswrapper[4573]: I1203 08:52:30.021449 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"a7a7a94f114c98ab77ca68e0f64ee624f3707f0f2fb9498f0fe627b281097c6f"} Dec 03 08:52:30 crc kubenswrapper[4573]: I1203 08:52:30.042904 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48186949-d112-4fb7-aa60-026d0176c73e" path="/var/lib/kubelet/pods/48186949-d112-4fb7-aa60-026d0176c73e/volumes" Dec 03 08:52:31 crc kubenswrapper[4573]: I1203 08:52:31.043747 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"9da33aae3c797fc1be8bf5d2340048c757d001091101648a29562399fe218985"} Dec 03 08:52:31 crc kubenswrapper[4573]: I1203 08:52:31.044201 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"f412bb7c05600902e1b567d6426e808b3a703127353cc47e1bf7ff5667e8bea4"} Dec 03 08:52:31 crc kubenswrapper[4573]: I1203 08:52:31.044214 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"777329a0eebbf330c4d6ed88362a1a6a6205ba01df419fbd82930bd1224361c7"} Dec 03 08:52:31 crc kubenswrapper[4573]: I1203 08:52:31.044225 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"6af6884e5beb91f72f42031a56958383a219feee21e145ffd4a67fb38729a2ab"} Dec 03 08:52:31 crc kubenswrapper[4573]: I1203 08:52:31.044235 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"1143d07859e797deb153da0ed267a79983d668c5f376d892db8cf0ee7a63d753"} Dec 03 08:52:31 crc kubenswrapper[4573]: I1203 08:52:31.044246 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"6fb806c7c28cbd3c30ea0925e953fbf46cb97ab1e2ce6fa453153a2ebbc241f5"} Dec 03 08:52:34 crc kubenswrapper[4573]: I1203 08:52:34.067143 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"1996c191f2bf9c758b20cbc644a94bdf13d1c01d55dce7932fdff10d29b790a5"} Dec 03 08:52:36 crc kubenswrapper[4573]: I1203 08:52:36.085180 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" event={"ID":"200c1dec-7901-4fe2-91d2-e4ad1ecbedf7","Type":"ContainerStarted","Data":"88b3ca9f862303e7e607b6a819b72077cf8e6997bcbf9209a9fde938f0180a32"} Dec 03 08:52:37 crc kubenswrapper[4573]: I1203 08:52:37.090985 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:37 crc kubenswrapper[4573]: I1203 08:52:37.091339 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:37 crc kubenswrapper[4573]: I1203 08:52:37.091355 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:37 crc kubenswrapper[4573]: I1203 08:52:37.129191 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" podStartSLOduration=9.129171768 podStartE2EDuration="9.129171768s" podCreationTimestamp="2025-12-03 08:52:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:52:37.122418815 +0000 UTC m=+877.690798084" watchObservedRunningTime="2025-12-03 08:52:37.129171768 +0000 UTC m=+877.697551037" Dec 03 08:52:37 crc kubenswrapper[4573]: I1203 08:52:37.134196 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:37 crc kubenswrapper[4573]: I1203 08:52:37.134567 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.294504 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dkl59"] Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.297562 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.336520 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkl59"] Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.401801 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdpsn\" (UniqueName: \"kubernetes.io/projected/5728f70f-14d6-4f4a-8db4-25660dd0e06d-kube-api-access-kdpsn\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.402083 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-catalog-content\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.402170 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-utilities\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.503884 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdpsn\" (UniqueName: \"kubernetes.io/projected/5728f70f-14d6-4f4a-8db4-25660dd0e06d-kube-api-access-kdpsn\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.504340 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-catalog-content\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.504515 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-utilities\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.504877 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-catalog-content\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.504893 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-utilities\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.525248 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdpsn\" (UniqueName: \"kubernetes.io/projected/5728f70f-14d6-4f4a-8db4-25660dd0e06d-kube-api-access-kdpsn\") pod \"redhat-marketplace-dkl59\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.620796 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:52:50 crc kubenswrapper[4573]: I1203 08:52:50.863683 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkl59"] Dec 03 08:52:50 crc kubenswrapper[4573]: W1203 08:52:50.871783 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5728f70f_14d6_4f4a_8db4_25660dd0e06d.slice/crio-4d07bdf97a78be535b16d5d5e3d84e15d5eb8b0c8b2603bc9504ad466789213c WatchSource:0}: Error finding container 4d07bdf97a78be535b16d5d5e3d84e15d5eb8b0c8b2603bc9504ad466789213c: Status 404 returned error can't find the container with id 4d07bdf97a78be535b16d5d5e3d84e15d5eb8b0c8b2603bc9504ad466789213c Dec 03 08:52:51 crc kubenswrapper[4573]: I1203 08:52:51.171413 4573 generic.go:334] "Generic (PLEG): container finished" podID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerID="414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638" exitCode=0 Dec 03 08:52:51 crc kubenswrapper[4573]: I1203 08:52:51.171499 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkl59" event={"ID":"5728f70f-14d6-4f4a-8db4-25660dd0e06d","Type":"ContainerDied","Data":"414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638"} Dec 03 08:52:51 crc kubenswrapper[4573]: I1203 08:52:51.171691 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkl59" event={"ID":"5728f70f-14d6-4f4a-8db4-25660dd0e06d","Type":"ContainerStarted","Data":"4d07bdf97a78be535b16d5d5e3d84e15d5eb8b0c8b2603bc9504ad466789213c"} Dec 03 08:52:52 crc kubenswrapper[4573]: I1203 08:52:52.181098 4573 generic.go:334] "Generic (PLEG): container finished" podID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerID="345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265" exitCode=0 Dec 03 08:52:52 crc kubenswrapper[4573]: I1203 08:52:52.181364 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkl59" event={"ID":"5728f70f-14d6-4f4a-8db4-25660dd0e06d","Type":"ContainerDied","Data":"345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265"} Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.077891 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cpfnk"] Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.079773 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.106361 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cpfnk"] Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.135948 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dptt\" (UniqueName: \"kubernetes.io/projected/388945d4-ebb9-4f1f-af34-6ff316f01b79-kube-api-access-7dptt\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.136062 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-utilities\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.136101 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-catalog-content\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.191928 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkl59" event={"ID":"5728f70f-14d6-4f4a-8db4-25660dd0e06d","Type":"ContainerStarted","Data":"5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f"} Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.222806 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dkl59" podStartSLOduration=1.823419404 podStartE2EDuration="3.222778276s" podCreationTimestamp="2025-12-03 08:52:50 +0000 UTC" firstStartedPulling="2025-12-03 08:52:51.173766171 +0000 UTC m=+891.742145430" lastFinishedPulling="2025-12-03 08:52:52.573125043 +0000 UTC m=+893.141504302" observedRunningTime="2025-12-03 08:52:53.218752067 +0000 UTC m=+893.787131346" watchObservedRunningTime="2025-12-03 08:52:53.222778276 +0000 UTC m=+893.791157535" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.237677 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dptt\" (UniqueName: \"kubernetes.io/projected/388945d4-ebb9-4f1f-af34-6ff316f01b79-kube-api-access-7dptt\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.237754 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-utilities\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.237804 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-catalog-content\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.238865 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-catalog-content\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.238965 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-utilities\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.264372 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dptt\" (UniqueName: \"kubernetes.io/projected/388945d4-ebb9-4f1f-af34-6ff316f01b79-kube-api-access-7dptt\") pod \"community-operators-cpfnk\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.399807 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:52:53 crc kubenswrapper[4573]: I1203 08:52:53.810808 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cpfnk"] Dec 03 08:52:53 crc kubenswrapper[4573]: W1203 08:52:53.824332 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod388945d4_ebb9_4f1f_af34_6ff316f01b79.slice/crio-f94deeba2e23a1bf1c0b2e76dd02a1d5affb142845a23c41815d9b4043627747 WatchSource:0}: Error finding container f94deeba2e23a1bf1c0b2e76dd02a1d5affb142845a23c41815d9b4043627747: Status 404 returned error can't find the container with id f94deeba2e23a1bf1c0b2e76dd02a1d5affb142845a23c41815d9b4043627747 Dec 03 08:52:54 crc kubenswrapper[4573]: I1203 08:52:54.199204 4573 generic.go:334] "Generic (PLEG): container finished" podID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerID="509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685" exitCode=0 Dec 03 08:52:54 crc kubenswrapper[4573]: I1203 08:52:54.199308 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpfnk" event={"ID":"388945d4-ebb9-4f1f-af34-6ff316f01b79","Type":"ContainerDied","Data":"509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685"} Dec 03 08:52:54 crc kubenswrapper[4573]: I1203 08:52:54.200668 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpfnk" event={"ID":"388945d4-ebb9-4f1f-af34-6ff316f01b79","Type":"ContainerStarted","Data":"f94deeba2e23a1bf1c0b2e76dd02a1d5affb142845a23c41815d9b4043627747"} Dec 03 08:52:55 crc kubenswrapper[4573]: I1203 08:52:55.209771 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpfnk" event={"ID":"388945d4-ebb9-4f1f-af34-6ff316f01b79","Type":"ContainerStarted","Data":"d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e"} Dec 03 08:52:56 crc kubenswrapper[4573]: I1203 08:52:56.218778 4573 generic.go:334] "Generic (PLEG): container finished" podID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerID="d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e" exitCode=0 Dec 03 08:52:56 crc kubenswrapper[4573]: I1203 08:52:56.218823 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpfnk" event={"ID":"388945d4-ebb9-4f1f-af34-6ff316f01b79","Type":"ContainerDied","Data":"d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e"} Dec 03 08:52:57 crc kubenswrapper[4573]: I1203 08:52:57.227957 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpfnk" event={"ID":"388945d4-ebb9-4f1f-af34-6ff316f01b79","Type":"ContainerStarted","Data":"7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e"} Dec 03 08:52:57 crc kubenswrapper[4573]: I1203 08:52:57.254198 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cpfnk" podStartSLOduration=1.7679053059999998 podStartE2EDuration="4.254167032s" podCreationTimestamp="2025-12-03 08:52:53 +0000 UTC" firstStartedPulling="2025-12-03 08:52:54.201131858 +0000 UTC m=+894.769511117" lastFinishedPulling="2025-12-03 08:52:56.687393544 +0000 UTC m=+897.255772843" observedRunningTime="2025-12-03 08:52:57.250028109 +0000 UTC m=+897.818407378" watchObservedRunningTime="2025-12-03 08:52:57.254167032 +0000 UTC m=+897.822546291" Dec 03 08:52:59 crc kubenswrapper[4573]: I1203 08:52:59.320887 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hb68s" Dec 03 08:53:00 crc kubenswrapper[4573]: I1203 08:53:00.621475 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:53:00 crc kubenswrapper[4573]: I1203 08:53:00.621586 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:53:00 crc kubenswrapper[4573]: I1203 08:53:00.670129 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:53:01 crc kubenswrapper[4573]: I1203 08:53:01.312613 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:53:01 crc kubenswrapper[4573]: I1203 08:53:01.379306 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkl59"] Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.267534 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dkl59" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerName="registry-server" containerID="cri-o://5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f" gracePeriod=2 Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.400118 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.400384 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.452359 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.633868 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.707071 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdpsn\" (UniqueName: \"kubernetes.io/projected/5728f70f-14d6-4f4a-8db4-25660dd0e06d-kube-api-access-kdpsn\") pod \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.707511 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-catalog-content\") pod \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.707762 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-utilities\") pod \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\" (UID: \"5728f70f-14d6-4f4a-8db4-25660dd0e06d\") " Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.709069 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-utilities" (OuterVolumeSpecName: "utilities") pod "5728f70f-14d6-4f4a-8db4-25660dd0e06d" (UID: "5728f70f-14d6-4f4a-8db4-25660dd0e06d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.714702 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5728f70f-14d6-4f4a-8db4-25660dd0e06d-kube-api-access-kdpsn" (OuterVolumeSpecName: "kube-api-access-kdpsn") pod "5728f70f-14d6-4f4a-8db4-25660dd0e06d" (UID: "5728f70f-14d6-4f4a-8db4-25660dd0e06d"). InnerVolumeSpecName "kube-api-access-kdpsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.732809 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5728f70f-14d6-4f4a-8db4-25660dd0e06d" (UID: "5728f70f-14d6-4f4a-8db4-25660dd0e06d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.809812 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.809860 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5728f70f-14d6-4f4a-8db4-25660dd0e06d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:03 crc kubenswrapper[4573]: I1203 08:53:03.809874 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdpsn\" (UniqueName: \"kubernetes.io/projected/5728f70f-14d6-4f4a-8db4-25660dd0e06d-kube-api-access-kdpsn\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.275820 4573 generic.go:334] "Generic (PLEG): container finished" podID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerID="5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f" exitCode=0 Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.275857 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkl59" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.275925 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkl59" event={"ID":"5728f70f-14d6-4f4a-8db4-25660dd0e06d","Type":"ContainerDied","Data":"5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f"} Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.276428 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkl59" event={"ID":"5728f70f-14d6-4f4a-8db4-25660dd0e06d","Type":"ContainerDied","Data":"4d07bdf97a78be535b16d5d5e3d84e15d5eb8b0c8b2603bc9504ad466789213c"} Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.276464 4573 scope.go:117] "RemoveContainer" containerID="5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.306914 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkl59"] Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.309009 4573 scope.go:117] "RemoveContainer" containerID="345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.311479 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkl59"] Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.328594 4573 scope.go:117] "RemoveContainer" containerID="414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.341624 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.349395 4573 scope.go:117] "RemoveContainer" containerID="5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f" Dec 03 08:53:04 crc kubenswrapper[4573]: E1203 08:53:04.350298 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f\": container with ID starting with 5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f not found: ID does not exist" containerID="5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.350359 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f"} err="failed to get container status \"5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f\": rpc error: code = NotFound desc = could not find container \"5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f\": container with ID starting with 5a46f83c6c389f69f8a48f94420987de0bb18dda7126d20fdc9adbe79903524f not found: ID does not exist" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.350401 4573 scope.go:117] "RemoveContainer" containerID="345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265" Dec 03 08:53:04 crc kubenswrapper[4573]: E1203 08:53:04.350989 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265\": container with ID starting with 345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265 not found: ID does not exist" containerID="345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.351059 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265"} err="failed to get container status \"345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265\": rpc error: code = NotFound desc = could not find container \"345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265\": container with ID starting with 345edfbfe42d206de115b18384493c2dc35d925d1f2c9ed4711875d3a336a265 not found: ID does not exist" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.351094 4573 scope.go:117] "RemoveContainer" containerID="414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638" Dec 03 08:53:04 crc kubenswrapper[4573]: E1203 08:53:04.351461 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638\": container with ID starting with 414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638 not found: ID does not exist" containerID="414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638" Dec 03 08:53:04 crc kubenswrapper[4573]: I1203 08:53:04.351484 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638"} err="failed to get container status \"414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638\": rpc error: code = NotFound desc = could not find container \"414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638\": container with ID starting with 414993c0cfdffdd869a5ad46a2a934cd4c4efa27ace763de369dc78173e07638 not found: ID does not exist" Dec 03 08:53:06 crc kubenswrapper[4573]: I1203 08:53:06.038553 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" path="/var/lib/kubelet/pods/5728f70f-14d6-4f4a-8db4-25660dd0e06d/volumes" Dec 03 08:53:06 crc kubenswrapper[4573]: I1203 08:53:06.710645 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cpfnk"] Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.296969 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cpfnk" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerName="registry-server" containerID="cri-o://7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e" gracePeriod=2 Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.625232 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.662677 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-utilities\") pod \"388945d4-ebb9-4f1f-af34-6ff316f01b79\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.663014 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dptt\" (UniqueName: \"kubernetes.io/projected/388945d4-ebb9-4f1f-af34-6ff316f01b79-kube-api-access-7dptt\") pod \"388945d4-ebb9-4f1f-af34-6ff316f01b79\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.663101 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-catalog-content\") pod \"388945d4-ebb9-4f1f-af34-6ff316f01b79\" (UID: \"388945d4-ebb9-4f1f-af34-6ff316f01b79\") " Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.664504 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-utilities" (OuterVolumeSpecName: "utilities") pod "388945d4-ebb9-4f1f-af34-6ff316f01b79" (UID: "388945d4-ebb9-4f1f-af34-6ff316f01b79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.670530 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/388945d4-ebb9-4f1f-af34-6ff316f01b79-kube-api-access-7dptt" (OuterVolumeSpecName: "kube-api-access-7dptt") pod "388945d4-ebb9-4f1f-af34-6ff316f01b79" (UID: "388945d4-ebb9-4f1f-af34-6ff316f01b79"). InnerVolumeSpecName "kube-api-access-7dptt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.717973 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "388945d4-ebb9-4f1f-af34-6ff316f01b79" (UID: "388945d4-ebb9-4f1f-af34-6ff316f01b79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.764825 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dptt\" (UniqueName: \"kubernetes.io/projected/388945d4-ebb9-4f1f-af34-6ff316f01b79-kube-api-access-7dptt\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.764892 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:07 crc kubenswrapper[4573]: I1203 08:53:07.764910 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/388945d4-ebb9-4f1f-af34-6ff316f01b79-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.303502 4573 generic.go:334] "Generic (PLEG): container finished" podID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerID="7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e" exitCode=0 Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.303560 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpfnk" event={"ID":"388945d4-ebb9-4f1f-af34-6ff316f01b79","Type":"ContainerDied","Data":"7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e"} Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.303600 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpfnk" event={"ID":"388945d4-ebb9-4f1f-af34-6ff316f01b79","Type":"ContainerDied","Data":"f94deeba2e23a1bf1c0b2e76dd02a1d5affb142845a23c41815d9b4043627747"} Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.303627 4573 scope.go:117] "RemoveContainer" containerID="7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.303775 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpfnk" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.326562 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cpfnk"] Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.329770 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cpfnk"] Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.353086 4573 scope.go:117] "RemoveContainer" containerID="d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.373749 4573 scope.go:117] "RemoveContainer" containerID="509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.394125 4573 scope.go:117] "RemoveContainer" containerID="7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e" Dec 03 08:53:08 crc kubenswrapper[4573]: E1203 08:53:08.394732 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e\": container with ID starting with 7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e not found: ID does not exist" containerID="7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.394796 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e"} err="failed to get container status \"7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e\": rpc error: code = NotFound desc = could not find container \"7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e\": container with ID starting with 7cfb02b7fb657e2448bba6c86d6bca568356f77ee3d5f6cb82e39aa6212b0a8e not found: ID does not exist" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.394838 4573 scope.go:117] "RemoveContainer" containerID="d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e" Dec 03 08:53:08 crc kubenswrapper[4573]: E1203 08:53:08.395361 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e\": container with ID starting with d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e not found: ID does not exist" containerID="d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.395416 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e"} err="failed to get container status \"d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e\": rpc error: code = NotFound desc = could not find container \"d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e\": container with ID starting with d398b742d4674cc8e6f8c6a0be9bd4d13734cfb64620524754119e2ba0e29d3e not found: ID does not exist" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.395451 4573 scope.go:117] "RemoveContainer" containerID="509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685" Dec 03 08:53:08 crc kubenswrapper[4573]: E1203 08:53:08.396114 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685\": container with ID starting with 509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685 not found: ID does not exist" containerID="509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685" Dec 03 08:53:08 crc kubenswrapper[4573]: I1203 08:53:08.396145 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685"} err="failed to get container status \"509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685\": rpc error: code = NotFound desc = could not find container \"509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685\": container with ID starting with 509ad7b50feca7710dc5346e83d310a6036cd370e43836c6b746330c546aa685 not found: ID does not exist" Dec 03 08:53:10 crc kubenswrapper[4573]: I1203 08:53:10.038589 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" path="/var/lib/kubelet/pods/388945d4-ebb9-4f1f-af34-6ff316f01b79/volumes" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.578560 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869"] Dec 03 08:53:13 crc kubenswrapper[4573]: E1203 08:53:13.579101 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerName="registry-server" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.579117 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerName="registry-server" Dec 03 08:53:13 crc kubenswrapper[4573]: E1203 08:53:13.579135 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerName="extract-content" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.579144 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerName="extract-content" Dec 03 08:53:13 crc kubenswrapper[4573]: E1203 08:53:13.579156 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerName="extract-utilities" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.579166 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerName="extract-utilities" Dec 03 08:53:13 crc kubenswrapper[4573]: E1203 08:53:13.579180 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerName="extract-utilities" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.579189 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerName="extract-utilities" Dec 03 08:53:13 crc kubenswrapper[4573]: E1203 08:53:13.579210 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerName="extract-content" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.579218 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerName="extract-content" Dec 03 08:53:13 crc kubenswrapper[4573]: E1203 08:53:13.579233 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerName="registry-server" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.579242 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerName="registry-server" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.579361 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="388945d4-ebb9-4f1f-af34-6ff316f01b79" containerName="registry-server" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.579375 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="5728f70f-14d6-4f4a-8db4-25660dd0e06d" containerName="registry-server" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.580335 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.583633 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.596386 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869"] Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.666394 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpjkm\" (UniqueName: \"kubernetes.io/projected/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-kube-api-access-xpjkm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.666486 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.666696 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.767979 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.768081 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpjkm\" (UniqueName: \"kubernetes.io/projected/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-kube-api-access-xpjkm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.768132 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.768596 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.768737 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.789103 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpjkm\" (UniqueName: \"kubernetes.io/projected/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-kube-api-access-xpjkm\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:13 crc kubenswrapper[4573]: I1203 08:53:13.904752 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:14 crc kubenswrapper[4573]: I1203 08:53:14.357787 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869"] Dec 03 08:53:15 crc kubenswrapper[4573]: I1203 08:53:15.348710 4573 generic.go:334] "Generic (PLEG): container finished" podID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerID="53d88320d3f5832b92eb87cd8cf3fa60be99c2a967e23b5f34844a17cf3f4734" exitCode=0 Dec 03 08:53:15 crc kubenswrapper[4573]: I1203 08:53:15.348783 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" event={"ID":"3e1c6c46-dab7-403c-afd3-3ff5469a2d36","Type":"ContainerDied","Data":"53d88320d3f5832b92eb87cd8cf3fa60be99c2a967e23b5f34844a17cf3f4734"} Dec 03 08:53:15 crc kubenswrapper[4573]: I1203 08:53:15.349037 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" event={"ID":"3e1c6c46-dab7-403c-afd3-3ff5469a2d36","Type":"ContainerStarted","Data":"69b5511dbec0871c4233faf1a4e8721291aa5af46b1e0d15e2b961d61bc9a162"} Dec 03 08:53:15 crc kubenswrapper[4573]: I1203 08:53:15.924616 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xlzzv"] Dec 03 08:53:15 crc kubenswrapper[4573]: I1203 08:53:15.926850 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:15 crc kubenswrapper[4573]: I1203 08:53:15.929212 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlzzv"] Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.005078 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgpql\" (UniqueName: \"kubernetes.io/projected/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-kube-api-access-fgpql\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.005189 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-catalog-content\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.005230 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-utilities\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.106255 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-catalog-content\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.106329 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-utilities\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.106389 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgpql\" (UniqueName: \"kubernetes.io/projected/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-kube-api-access-fgpql\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.107377 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-utilities\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.107467 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-catalog-content\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.148859 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgpql\" (UniqueName: \"kubernetes.io/projected/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-kube-api-access-fgpql\") pod \"redhat-operators-xlzzv\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.245475 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:16 crc kubenswrapper[4573]: I1203 08:53:16.538717 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xlzzv"] Dec 03 08:53:16 crc kubenswrapper[4573]: W1203 08:53:16.600748 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ec8d9d8_d19c_4e2f_ba1d_31aaf7839966.slice/crio-01b1e89312af9bd759b5a15b94c7f6eb400ecaa0871455f016988f8ba08546a3 WatchSource:0}: Error finding container 01b1e89312af9bd759b5a15b94c7f6eb400ecaa0871455f016988f8ba08546a3: Status 404 returned error can't find the container with id 01b1e89312af9bd759b5a15b94c7f6eb400ecaa0871455f016988f8ba08546a3 Dec 03 08:53:17 crc kubenswrapper[4573]: I1203 08:53:17.369772 4573 generic.go:334] "Generic (PLEG): container finished" podID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerID="a51b385b6809123c6749caa1b0833d30a6e3ef539b83fcce1e8a00c8b9e4c7cd" exitCode=0 Dec 03 08:53:17 crc kubenswrapper[4573]: I1203 08:53:17.369878 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" event={"ID":"3e1c6c46-dab7-403c-afd3-3ff5469a2d36","Type":"ContainerDied","Data":"a51b385b6809123c6749caa1b0833d30a6e3ef539b83fcce1e8a00c8b9e4c7cd"} Dec 03 08:53:17 crc kubenswrapper[4573]: I1203 08:53:17.373899 4573 generic.go:334] "Generic (PLEG): container finished" podID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerID="c481ae59d24cdce12f4666d1632702c6760065c2a60fd88836c873deced79bc2" exitCode=0 Dec 03 08:53:17 crc kubenswrapper[4573]: I1203 08:53:17.373943 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlzzv" event={"ID":"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966","Type":"ContainerDied","Data":"c481ae59d24cdce12f4666d1632702c6760065c2a60fd88836c873deced79bc2"} Dec 03 08:53:17 crc kubenswrapper[4573]: I1203 08:53:17.373996 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlzzv" event={"ID":"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966","Type":"ContainerStarted","Data":"01b1e89312af9bd759b5a15b94c7f6eb400ecaa0871455f016988f8ba08546a3"} Dec 03 08:53:18 crc kubenswrapper[4573]: I1203 08:53:18.389608 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlzzv" event={"ID":"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966","Type":"ContainerStarted","Data":"8e2d5db8d77909d812ea7ba8d1d154ba3559ef922a4d0251ecc43a6b8a7f8376"} Dec 03 08:53:18 crc kubenswrapper[4573]: I1203 08:53:18.395285 4573 generic.go:334] "Generic (PLEG): container finished" podID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerID="db9afb0f1eb5fda147e2ad5342737a0863d0e64bc2f90786fd8d938811d27702" exitCode=0 Dec 03 08:53:18 crc kubenswrapper[4573]: I1203 08:53:18.395344 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" event={"ID":"3e1c6c46-dab7-403c-afd3-3ff5469a2d36","Type":"ContainerDied","Data":"db9afb0f1eb5fda147e2ad5342737a0863d0e64bc2f90786fd8d938811d27702"} Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.284875 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.362491 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-bundle\") pod \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.362624 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpjkm\" (UniqueName: \"kubernetes.io/projected/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-kube-api-access-xpjkm\") pod \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.362688 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-util\") pod \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\" (UID: \"3e1c6c46-dab7-403c-afd3-3ff5469a2d36\") " Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.363639 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-bundle" (OuterVolumeSpecName: "bundle") pod "3e1c6c46-dab7-403c-afd3-3ff5469a2d36" (UID: "3e1c6c46-dab7-403c-afd3-3ff5469a2d36"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.376702 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-util" (OuterVolumeSpecName: "util") pod "3e1c6c46-dab7-403c-afd3-3ff5469a2d36" (UID: "3e1c6c46-dab7-403c-afd3-3ff5469a2d36"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.377177 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-kube-api-access-xpjkm" (OuterVolumeSpecName: "kube-api-access-xpjkm") pod "3e1c6c46-dab7-403c-afd3-3ff5469a2d36" (UID: "3e1c6c46-dab7-403c-afd3-3ff5469a2d36"). InnerVolumeSpecName "kube-api-access-xpjkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.464993 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpjkm\" (UniqueName: \"kubernetes.io/projected/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-kube-api-access-xpjkm\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.465039 4573 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-util\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.465070 4573 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3e1c6c46-dab7-403c-afd3-3ff5469a2d36-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.781186 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" event={"ID":"3e1c6c46-dab7-403c-afd3-3ff5469a2d36","Type":"ContainerDied","Data":"69b5511dbec0871c4233faf1a4e8721291aa5af46b1e0d15e2b961d61bc9a162"} Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.781578 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69b5511dbec0871c4233faf1a4e8721291aa5af46b1e0d15e2b961d61bc9a162" Dec 03 08:53:20 crc kubenswrapper[4573]: I1203 08:53:20.781276 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869" Dec 03 08:53:21 crc kubenswrapper[4573]: I1203 08:53:21.793388 4573 generic.go:334] "Generic (PLEG): container finished" podID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerID="8e2d5db8d77909d812ea7ba8d1d154ba3559ef922a4d0251ecc43a6b8a7f8376" exitCode=0 Dec 03 08:53:21 crc kubenswrapper[4573]: I1203 08:53:21.793464 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlzzv" event={"ID":"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966","Type":"ContainerDied","Data":"8e2d5db8d77909d812ea7ba8d1d154ba3559ef922a4d0251ecc43a6b8a7f8376"} Dec 03 08:53:22 crc kubenswrapper[4573]: I1203 08:53:22.801958 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlzzv" event={"ID":"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966","Type":"ContainerStarted","Data":"faa36b4497d830e762ff75f1ec4c0183d436dd1b2fb3ea7e105092b35fd7bab3"} Dec 03 08:53:22 crc kubenswrapper[4573]: I1203 08:53:22.825652 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xlzzv" podStartSLOduration=2.985730296 podStartE2EDuration="7.825628663s" podCreationTimestamp="2025-12-03 08:53:15 +0000 UTC" firstStartedPulling="2025-12-03 08:53:17.375140306 +0000 UTC m=+917.943519565" lastFinishedPulling="2025-12-03 08:53:22.215038673 +0000 UTC m=+922.783417932" observedRunningTime="2025-12-03 08:53:22.8203655 +0000 UTC m=+923.388744769" watchObservedRunningTime="2025-12-03 08:53:22.825628663 +0000 UTC m=+923.394007922" Dec 03 08:53:24 crc kubenswrapper[4573]: I1203 08:53:24.931611 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4"] Dec 03 08:53:24 crc kubenswrapper[4573]: E1203 08:53:24.932441 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerName="util" Dec 03 08:53:24 crc kubenswrapper[4573]: I1203 08:53:24.932462 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerName="util" Dec 03 08:53:24 crc kubenswrapper[4573]: E1203 08:53:24.932476 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerName="pull" Dec 03 08:53:24 crc kubenswrapper[4573]: I1203 08:53:24.932483 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerName="pull" Dec 03 08:53:24 crc kubenswrapper[4573]: E1203 08:53:24.932497 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerName="extract" Dec 03 08:53:24 crc kubenswrapper[4573]: I1203 08:53:24.932503 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerName="extract" Dec 03 08:53:24 crc kubenswrapper[4573]: I1203 08:53:24.932615 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e1c6c46-dab7-403c-afd3-3ff5469a2d36" containerName="extract" Dec 03 08:53:24 crc kubenswrapper[4573]: I1203 08:53:24.933200 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" Dec 03 08:53:24 crc kubenswrapper[4573]: W1203 08:53:24.937975 4573 reflector.go:561] object-"openshift-nmstate"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Dec 03 08:53:24 crc kubenswrapper[4573]: E1203 08:53:24.939414 4573 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 08:53:24 crc kubenswrapper[4573]: W1203 08:53:24.938005 4573 reflector.go:561] object-"openshift-nmstate"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Dec 03 08:53:24 crc kubenswrapper[4573]: E1203 08:53:24.939916 4573 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 08:53:24 crc kubenswrapper[4573]: W1203 08:53:24.942121 4573 reflector.go:561] object-"openshift-nmstate"/"nmstate-operator-dockercfg-87nhd": failed to list *v1.Secret: secrets "nmstate-operator-dockercfg-87nhd" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-nmstate": no relationship found between node 'crc' and this object Dec 03 08:53:24 crc kubenswrapper[4573]: E1203 08:53:24.942182 4573 reflector.go:158] "Unhandled Error" err="object-\"openshift-nmstate\"/\"nmstate-operator-dockercfg-87nhd\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"nmstate-operator-dockercfg-87nhd\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-nmstate\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 08:53:24 crc kubenswrapper[4573]: I1203 08:53:24.953072 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jbch\" (UniqueName: \"kubernetes.io/projected/d8f81a47-9022-45cf-93e5-8989ee9121ce-kube-api-access-9jbch\") pod \"nmstate-operator-5b5b58f5c8-j7vz4\" (UID: \"d8f81a47-9022-45cf-93e5-8989ee9121ce\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" Dec 03 08:53:24 crc kubenswrapper[4573]: I1203 08:53:24.965998 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4"] Dec 03 08:53:25 crc kubenswrapper[4573]: I1203 08:53:25.054976 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jbch\" (UniqueName: \"kubernetes.io/projected/d8f81a47-9022-45cf-93e5-8989ee9121ce-kube-api-access-9jbch\") pod \"nmstate-operator-5b5b58f5c8-j7vz4\" (UID: \"d8f81a47-9022-45cf-93e5-8989ee9121ce\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" Dec 03 08:53:26 crc kubenswrapper[4573]: E1203 08:53:26.075377 4573 projected.go:288] Couldn't get configMap openshift-nmstate/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.246761 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.246869 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.267709 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-87nhd" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.335303 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.427587 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 08:53:26 crc kubenswrapper[4573]: E1203 08:53:26.436327 4573 projected.go:194] Error preparing data for projected volume kube-api-access-9jbch for pod openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4: failed to sync configmap cache: timed out waiting for the condition Dec 03 08:53:26 crc kubenswrapper[4573]: E1203 08:53:26.436512 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d8f81a47-9022-45cf-93e5-8989ee9121ce-kube-api-access-9jbch podName:d8f81a47-9022-45cf-93e5-8989ee9121ce nodeName:}" failed. No retries permitted until 2025-12-03 08:53:26.936477019 +0000 UTC m=+927.504856278 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9jbch" (UniqueName: "kubernetes.io/projected/d8f81a47-9022-45cf-93e5-8989ee9121ce-kube-api-access-9jbch") pod "nmstate-operator-5b5b58f5c8-j7vz4" (UID: "d8f81a47-9022-45cf-93e5-8989ee9121ce") : failed to sync configmap cache: timed out waiting for the condition Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.521435 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lc8zc"] Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.524421 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.560976 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lc8zc"] Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.578386 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-catalog-content\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.578749 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6x8v\" (UniqueName: \"kubernetes.io/projected/50c256a6-8345-4a1b-96f7-8475dcc28c47-kube-api-access-x6x8v\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.578914 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-utilities\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.680196 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6x8v\" (UniqueName: \"kubernetes.io/projected/50c256a6-8345-4a1b-96f7-8475dcc28c47-kube-api-access-x6x8v\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.680321 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-utilities\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.680385 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-catalog-content\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.681035 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-catalog-content\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.681195 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-utilities\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.716633 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6x8v\" (UniqueName: \"kubernetes.io/projected/50c256a6-8345-4a1b-96f7-8475dcc28c47-kube-api-access-x6x8v\") pod \"certified-operators-lc8zc\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.846735 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:26 crc kubenswrapper[4573]: I1203 08:53:26.986018 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jbch\" (UniqueName: \"kubernetes.io/projected/d8f81a47-9022-45cf-93e5-8989ee9121ce-kube-api-access-9jbch\") pod \"nmstate-operator-5b5b58f5c8-j7vz4\" (UID: \"d8f81a47-9022-45cf-93e5-8989ee9121ce\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" Dec 03 08:53:27 crc kubenswrapper[4573]: I1203 08:53:27.033451 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jbch\" (UniqueName: \"kubernetes.io/projected/d8f81a47-9022-45cf-93e5-8989ee9121ce-kube-api-access-9jbch\") pod \"nmstate-operator-5b5b58f5c8-j7vz4\" (UID: \"d8f81a47-9022-45cf-93e5-8989ee9121ce\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" Dec 03 08:53:27 crc kubenswrapper[4573]: I1203 08:53:27.055374 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" Dec 03 08:53:27 crc kubenswrapper[4573]: I1203 08:53:27.329294 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xlzzv" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="registry-server" probeResult="failure" output=< Dec 03 08:53:27 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 08:53:27 crc kubenswrapper[4573]: > Dec 03 08:53:27 crc kubenswrapper[4573]: I1203 08:53:27.534452 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lc8zc"] Dec 03 08:53:28 crc kubenswrapper[4573]: I1203 08:53:27.837157 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc8zc" event={"ID":"50c256a6-8345-4a1b-96f7-8475dcc28c47","Type":"ContainerStarted","Data":"3de93eb9cc743fd43780bb53d327ff01fd9b409058076e742551b6a17192fc45"} Dec 03 08:53:28 crc kubenswrapper[4573]: I1203 08:53:28.199722 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4"] Dec 03 08:53:28 crc kubenswrapper[4573]: I1203 08:53:28.846308 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" event={"ID":"d8f81a47-9022-45cf-93e5-8989ee9121ce","Type":"ContainerStarted","Data":"5f6733b96f6e80134dc6231ab388ca8ce9e0d8b32816ecedba0f0dc661b13a4a"} Dec 03 08:53:29 crc kubenswrapper[4573]: I1203 08:53:29.904956 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc8zc" event={"ID":"50c256a6-8345-4a1b-96f7-8475dcc28c47","Type":"ContainerStarted","Data":"d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8"} Dec 03 08:53:31 crc kubenswrapper[4573]: I1203 08:53:31.017892 4573 generic.go:334] "Generic (PLEG): container finished" podID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerID="d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8" exitCode=0 Dec 03 08:53:31 crc kubenswrapper[4573]: I1203 08:53:31.017960 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc8zc" event={"ID":"50c256a6-8345-4a1b-96f7-8475dcc28c47","Type":"ContainerDied","Data":"d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8"} Dec 03 08:53:32 crc kubenswrapper[4573]: I1203 08:53:32.025782 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" event={"ID":"d8f81a47-9022-45cf-93e5-8989ee9121ce","Type":"ContainerStarted","Data":"603536a204c1da5e539d36d6a0d812719fc37acf510f4c40e3c03ab6fd7396ec"} Dec 03 08:53:32 crc kubenswrapper[4573]: I1203 08:53:32.045269 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-j7vz4" podStartSLOduration=4.528184466 podStartE2EDuration="8.045254212s" podCreationTimestamp="2025-12-03 08:53:24 +0000 UTC" firstStartedPulling="2025-12-03 08:53:28.214183659 +0000 UTC m=+928.782562918" lastFinishedPulling="2025-12-03 08:53:31.731253405 +0000 UTC m=+932.299632664" observedRunningTime="2025-12-03 08:53:32.044259824 +0000 UTC m=+932.612639083" watchObservedRunningTime="2025-12-03 08:53:32.045254212 +0000 UTC m=+932.613633471" Dec 03 08:53:33 crc kubenswrapper[4573]: I1203 08:53:33.036805 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc8zc" event={"ID":"50c256a6-8345-4a1b-96f7-8475dcc28c47","Type":"ContainerStarted","Data":"024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986"} Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.007245 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.008931 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.014432 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-k54wc" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.040150 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.047564 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.048421 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.076700 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6mwnn"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.077775 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.078377 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.096965 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.212857 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz9rz\" (UniqueName: \"kubernetes.io/projected/483c74ea-46d0-42e4-84c3-35ed45a0507e-kube-api-access-lz9rz\") pod \"nmstate-metrics-7f946cbc9-g2mdj\" (UID: \"483c74ea-46d0-42e4-84c3-35ed45a0507e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.213213 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8svch\" (UniqueName: \"kubernetes.io/projected/358ba50e-2a33-45a2-b1d4-b606638246d5-kube-api-access-8svch\") pod \"nmstate-webhook-5f6d4c5ccb-hlbnj\" (UID: \"358ba50e-2a33-45a2-b1d4-b606638246d5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.213355 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-dbus-socket\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.213590 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-nmstate-lock\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.213806 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/358ba50e-2a33-45a2-b1d4-b606638246d5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-hlbnj\" (UID: \"358ba50e-2a33-45a2-b1d4-b606638246d5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.213966 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78nd4\" (UniqueName: \"kubernetes.io/projected/72904583-c577-4d6a-91e1-3ec1d51ec7c2-kube-api-access-78nd4\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.214123 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-ovs-socket\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.287887 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.288909 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.299629 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.302151 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.302850 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.322586 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-2t82z" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.327258 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz9rz\" (UniqueName: \"kubernetes.io/projected/483c74ea-46d0-42e4-84c3-35ed45a0507e-kube-api-access-lz9rz\") pod \"nmstate-metrics-7f946cbc9-g2mdj\" (UID: \"483c74ea-46d0-42e4-84c3-35ed45a0507e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.327309 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8svch\" (UniqueName: \"kubernetes.io/projected/358ba50e-2a33-45a2-b1d4-b606638246d5-kube-api-access-8svch\") pod \"nmstate-webhook-5f6d4c5ccb-hlbnj\" (UID: \"358ba50e-2a33-45a2-b1d4-b606638246d5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.327339 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-dbus-socket\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.327396 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-nmstate-lock\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.327422 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/358ba50e-2a33-45a2-b1d4-b606638246d5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-hlbnj\" (UID: \"358ba50e-2a33-45a2-b1d4-b606638246d5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.327451 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78nd4\" (UniqueName: \"kubernetes.io/projected/72904583-c577-4d6a-91e1-3ec1d51ec7c2-kube-api-access-78nd4\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.327474 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-ovs-socket\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.327884 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-dbus-socket\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.330393 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-nmstate-lock\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.332347 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/72904583-c577-4d6a-91e1-3ec1d51ec7c2-ovs-socket\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.361387 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/358ba50e-2a33-45a2-b1d4-b606638246d5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-hlbnj\" (UID: \"358ba50e-2a33-45a2-b1d4-b606638246d5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.362702 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78nd4\" (UniqueName: \"kubernetes.io/projected/72904583-c577-4d6a-91e1-3ec1d51ec7c2-kube-api-access-78nd4\") pod \"nmstate-handler-6mwnn\" (UID: \"72904583-c577-4d6a-91e1-3ec1d51ec7c2\") " pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.364215 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz9rz\" (UniqueName: \"kubernetes.io/projected/483c74ea-46d0-42e4-84c3-35ed45a0507e-kube-api-access-lz9rz\") pod \"nmstate-metrics-7f946cbc9-g2mdj\" (UID: \"483c74ea-46d0-42e4-84c3-35ed45a0507e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.393721 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8svch\" (UniqueName: \"kubernetes.io/projected/358ba50e-2a33-45a2-b1d4-b606638246d5-kube-api-access-8svch\") pod \"nmstate-webhook-5f6d4c5ccb-hlbnj\" (UID: \"358ba50e-2a33-45a2-b1d4-b606638246d5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.403683 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.428986 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vlnz\" (UniqueName: \"kubernetes.io/projected/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-kube-api-access-7vlnz\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.429086 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.429365 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: W1203 08:53:34.444622 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72904583_c577_4d6a_91e1_3ec1d51ec7c2.slice/crio-869a7f9d07c6ff5ab1914829cd7052fc50575404725ece1f1cac92d821f110e4 WatchSource:0}: Error finding container 869a7f9d07c6ff5ab1914829cd7052fc50575404725ece1f1cac92d821f110e4: Status 404 returned error can't find the container with id 869a7f9d07c6ff5ab1914829cd7052fc50575404725ece1f1cac92d821f110e4 Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.532934 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6f5d5cf467-m4grv"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.533869 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.535786 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/92a87bd3-653e-40bd-9e75-cf06885f4347-console-oauth-config\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.535821 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-trusted-ca-bundle\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.535854 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.535884 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-oauth-serving-cert\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.535906 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/92a87bd3-653e-40bd-9e75-cf06885f4347-console-serving-cert\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.535929 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.535955 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-console-config\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.535978 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-service-ca\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.536015 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vlnz\" (UniqueName: \"kubernetes.io/projected/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-kube-api-access-7vlnz\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.536033 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph6wm\" (UniqueName: \"kubernetes.io/projected/92a87bd3-653e-40bd-9e75-cf06885f4347-kube-api-access-ph6wm\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.537064 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: E1203 08:53:34.537169 4573 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 03 08:53:34 crc kubenswrapper[4573]: E1203 08:53:34.537228 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert podName:7006af64-f3a2-4410-84e4-f6dfd3bcf0ae nodeName:}" failed. No retries permitted until 2025-12-03 08:53:35.037207312 +0000 UTC m=+935.605586571 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-9wr28" (UID: "7006af64-f3a2-4410-84e4-f6dfd3bcf0ae") : secret "plugin-serving-cert" not found Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.560825 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f5d5cf467-m4grv"] Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.573472 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vlnz\" (UniqueName: \"kubernetes.io/projected/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-kube-api-access-7vlnz\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.629175 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.637068 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-oauth-serving-cert\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.637153 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/92a87bd3-653e-40bd-9e75-cf06885f4347-console-serving-cert\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.637229 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-console-config\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.637264 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-service-ca\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.637318 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph6wm\" (UniqueName: \"kubernetes.io/projected/92a87bd3-653e-40bd-9e75-cf06885f4347-kube-api-access-ph6wm\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.637363 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/92a87bd3-653e-40bd-9e75-cf06885f4347-console-oauth-config\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.637388 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-trusted-ca-bundle\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.638548 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-service-ca\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.638553 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-oauth-serving-cert\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.638933 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-trusted-ca-bundle\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.639404 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/92a87bd3-653e-40bd-9e75-cf06885f4347-console-config\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.643170 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/92a87bd3-653e-40bd-9e75-cf06885f4347-console-oauth-config\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.644328 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/92a87bd3-653e-40bd-9e75-cf06885f4347-console-serving-cert\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.661109 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph6wm\" (UniqueName: \"kubernetes.io/projected/92a87bd3-653e-40bd-9e75-cf06885f4347-kube-api-access-ph6wm\") pod \"console-6f5d5cf467-m4grv\" (UID: \"92a87bd3-653e-40bd-9e75-cf06885f4347\") " pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.680502 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:34 crc kubenswrapper[4573]: I1203 08:53:34.856525 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:35 crc kubenswrapper[4573]: I1203 08:53:35.053735 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:35 crc kubenswrapper[4573]: E1203 08:53:35.053953 4573 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 03 08:53:35 crc kubenswrapper[4573]: E1203 08:53:35.054026 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert podName:7006af64-f3a2-4410-84e4-f6dfd3bcf0ae nodeName:}" failed. No retries permitted until 2025-12-03 08:53:36.05400527 +0000 UTC m=+936.622384529 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-9wr28" (UID: "7006af64-f3a2-4410-84e4-f6dfd3bcf0ae") : secret "plugin-serving-cert" not found Dec 03 08:53:35 crc kubenswrapper[4573]: I1203 08:53:35.087992 4573 generic.go:334] "Generic (PLEG): container finished" podID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerID="024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986" exitCode=0 Dec 03 08:53:35 crc kubenswrapper[4573]: I1203 08:53:35.088123 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc8zc" event={"ID":"50c256a6-8345-4a1b-96f7-8475dcc28c47","Type":"ContainerDied","Data":"024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986"} Dec 03 08:53:35 crc kubenswrapper[4573]: I1203 08:53:35.106205 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6mwnn" event={"ID":"72904583-c577-4d6a-91e1-3ec1d51ec7c2","Type":"ContainerStarted","Data":"869a7f9d07c6ff5ab1914829cd7052fc50575404725ece1f1cac92d821f110e4"} Dec 03 08:53:35 crc kubenswrapper[4573]: I1203 08:53:35.535278 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj"] Dec 03 08:53:35 crc kubenswrapper[4573]: I1203 08:53:35.625807 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj"] Dec 03 08:53:35 crc kubenswrapper[4573]: W1203 08:53:35.647892 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod358ba50e_2a33_45a2_b1d4_b606638246d5.slice/crio-246b3ffef5101c721f8ffc3c6ac651b4441925e6aca65562733df4b564751ed8 WatchSource:0}: Error finding container 246b3ffef5101c721f8ffc3c6ac651b4441925e6aca65562733df4b564751ed8: Status 404 returned error can't find the container with id 246b3ffef5101c721f8ffc3c6ac651b4441925e6aca65562733df4b564751ed8 Dec 03 08:53:35 crc kubenswrapper[4573]: I1203 08:53:35.958733 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f5d5cf467-m4grv"] Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.075360 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.082909 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7006af64-f3a2-4410-84e4-f6dfd3bcf0ae-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-9wr28\" (UID: \"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.107554 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.126095 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f5d5cf467-m4grv" event={"ID":"92a87bd3-653e-40bd-9e75-cf06885f4347","Type":"ContainerStarted","Data":"d680945ac080c58a50c687386c398f0d139f8051a6d22adfe92f020c744bd8a5"} Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.129776 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" event={"ID":"483c74ea-46d0-42e4-84c3-35ed45a0507e","Type":"ContainerStarted","Data":"dbe0b839aac9ddf0158266b2f040930bee4c6f88b5f6b743e01a811f1adb3188"} Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.131316 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" event={"ID":"358ba50e-2a33-45a2-b1d4-b606638246d5","Type":"ContainerStarted","Data":"246b3ffef5101c721f8ffc3c6ac651b4441925e6aca65562733df4b564751ed8"} Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.316331 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.366792 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:36 crc kubenswrapper[4573]: I1203 08:53:36.431764 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28"] Dec 03 08:53:37 crc kubenswrapper[4573]: I1203 08:53:37.161475 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc8zc" event={"ID":"50c256a6-8345-4a1b-96f7-8475dcc28c47","Type":"ContainerStarted","Data":"b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe"} Dec 03 08:53:37 crc kubenswrapper[4573]: I1203 08:53:37.174580 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f5d5cf467-m4grv" event={"ID":"92a87bd3-653e-40bd-9e75-cf06885f4347","Type":"ContainerStarted","Data":"c063376814b8ba8b06e73817cdbcc2db8c15cea0cc6b8a93813d5d92a36150ae"} Dec 03 08:53:37 crc kubenswrapper[4573]: I1203 08:53:37.178114 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" event={"ID":"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae","Type":"ContainerStarted","Data":"40ee3024ccf873e140108370989c4ff4eabd2ab6b801a5e5d95f19365689370f"} Dec 03 08:53:37 crc kubenswrapper[4573]: I1203 08:53:37.188568 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lc8zc" podStartSLOduration=6.840709734 podStartE2EDuration="11.188547123s" podCreationTimestamp="2025-12-03 08:53:26 +0000 UTC" firstStartedPulling="2025-12-03 08:53:31.666263055 +0000 UTC m=+932.234642314" lastFinishedPulling="2025-12-03 08:53:36.014100444 +0000 UTC m=+936.582479703" observedRunningTime="2025-12-03 08:53:37.186700562 +0000 UTC m=+937.755079831" watchObservedRunningTime="2025-12-03 08:53:37.188547123 +0000 UTC m=+937.756926382" Dec 03 08:53:37 crc kubenswrapper[4573]: I1203 08:53:37.221896 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6f5d5cf467-m4grv" podStartSLOduration=3.22186718 podStartE2EDuration="3.22186718s" podCreationTimestamp="2025-12-03 08:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:53:37.221412507 +0000 UTC m=+937.789791786" watchObservedRunningTime="2025-12-03 08:53:37.22186718 +0000 UTC m=+937.790246439" Dec 03 08:53:38 crc kubenswrapper[4573]: I1203 08:53:38.722876 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlzzv"] Dec 03 08:53:38 crc kubenswrapper[4573]: I1203 08:53:38.723204 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xlzzv" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="registry-server" containerID="cri-o://faa36b4497d830e762ff75f1ec4c0183d436dd1b2fb3ea7e105092b35fd7bab3" gracePeriod=2 Dec 03 08:53:39 crc kubenswrapper[4573]: I1203 08:53:39.198670 4573 generic.go:334] "Generic (PLEG): container finished" podID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerID="faa36b4497d830e762ff75f1ec4c0183d436dd1b2fb3ea7e105092b35fd7bab3" exitCode=0 Dec 03 08:53:39 crc kubenswrapper[4573]: I1203 08:53:39.198745 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlzzv" event={"ID":"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966","Type":"ContainerDied","Data":"faa36b4497d830e762ff75f1ec4c0183d436dd1b2fb3ea7e105092b35fd7bab3"} Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.774835 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.777449 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgpql\" (UniqueName: \"kubernetes.io/projected/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-kube-api-access-fgpql\") pod \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.777540 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-catalog-content\") pod \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.777583 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-utilities\") pod \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\" (UID: \"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966\") " Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.778868 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-utilities" (OuterVolumeSpecName: "utilities") pod "6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" (UID: "6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.791414 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-kube-api-access-fgpql" (OuterVolumeSpecName: "kube-api-access-fgpql") pod "6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" (UID: "6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966"). InnerVolumeSpecName "kube-api-access-fgpql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.881192 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgpql\" (UniqueName: \"kubernetes.io/projected/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-kube-api-access-fgpql\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.881595 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.919172 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" (UID: "6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:41 crc kubenswrapper[4573]: I1203 08:53:41.983422 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.283735 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" event={"ID":"7006af64-f3a2-4410-84e4-f6dfd3bcf0ae","Type":"ContainerStarted","Data":"fa3b4a2c532679f845fabe8c5ceea08bdccdab8e1ed18376a0304205593e0386"} Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.287018 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" event={"ID":"483c74ea-46d0-42e4-84c3-35ed45a0507e","Type":"ContainerStarted","Data":"bafbd9133b46b7593563ddec53febcb280d0e182249d6ac7185b0a9d2aecb732"} Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.289692 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" event={"ID":"358ba50e-2a33-45a2-b1d4-b606638246d5","Type":"ContainerStarted","Data":"7255d72d2a3ff81412bafec3ea85b7d47a2474a082c6efcdcbabe7f7e1c891cc"} Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.290354 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.292208 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6mwnn" event={"ID":"72904583-c577-4d6a-91e1-3ec1d51ec7c2","Type":"ContainerStarted","Data":"b97525624bbc4dc0c278176e3b946c162531ce6ec1cec17667cce5c12dd59ccd"} Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.292664 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.300764 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xlzzv" event={"ID":"6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966","Type":"ContainerDied","Data":"01b1e89312af9bd759b5a15b94c7f6eb400ecaa0871455f016988f8ba08546a3"} Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.300833 4573 scope.go:117] "RemoveContainer" containerID="faa36b4497d830e762ff75f1ec4c0183d436dd1b2fb3ea7e105092b35fd7bab3" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.300866 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xlzzv" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.315925 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-9wr28" podStartSLOduration=2.887800577 podStartE2EDuration="8.31590117s" podCreationTimestamp="2025-12-03 08:53:34 +0000 UTC" firstStartedPulling="2025-12-03 08:53:36.443887463 +0000 UTC m=+937.012266722" lastFinishedPulling="2025-12-03 08:53:41.871988056 +0000 UTC m=+942.440367315" observedRunningTime="2025-12-03 08:53:42.310698818 +0000 UTC m=+942.879078077" watchObservedRunningTime="2025-12-03 08:53:42.31590117 +0000 UTC m=+942.884280439" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.372042 4573 scope.go:117] "RemoveContainer" containerID="8e2d5db8d77909d812ea7ba8d1d154ba3559ef922a4d0251ecc43a6b8a7f8376" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.374820 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" podStartSLOduration=2.131833049 podStartE2EDuration="8.374785853s" podCreationTimestamp="2025-12-03 08:53:34 +0000 UTC" firstStartedPulling="2025-12-03 08:53:35.650878397 +0000 UTC m=+936.219257656" lastFinishedPulling="2025-12-03 08:53:41.893831191 +0000 UTC m=+942.462210460" observedRunningTime="2025-12-03 08:53:42.371710019 +0000 UTC m=+942.940089278" watchObservedRunningTime="2025-12-03 08:53:42.374785853 +0000 UTC m=+942.943165112" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.401340 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6mwnn" podStartSLOduration=0.947730217 podStartE2EDuration="8.401324755s" podCreationTimestamp="2025-12-03 08:53:34 +0000 UTC" firstStartedPulling="2025-12-03 08:53:34.448832207 +0000 UTC m=+935.017211466" lastFinishedPulling="2025-12-03 08:53:41.902426745 +0000 UTC m=+942.470806004" observedRunningTime="2025-12-03 08:53:42.398618131 +0000 UTC m=+942.966997390" watchObservedRunningTime="2025-12-03 08:53:42.401324755 +0000 UTC m=+942.969704014" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.403101 4573 scope.go:117] "RemoveContainer" containerID="c481ae59d24cdce12f4666d1632702c6760065c2a60fd88836c873deced79bc2" Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.425983 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xlzzv"] Dec 03 08:53:42 crc kubenswrapper[4573]: I1203 08:53:42.439872 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xlzzv"] Dec 03 08:53:44 crc kubenswrapper[4573]: I1203 08:53:44.039932 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" path="/var/lib/kubelet/pods/6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966/volumes" Dec 03 08:53:44 crc kubenswrapper[4573]: I1203 08:53:44.857964 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:44 crc kubenswrapper[4573]: I1203 08:53:44.858038 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:44 crc kubenswrapper[4573]: I1203 08:53:44.864914 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:45 crc kubenswrapper[4573]: I1203 08:53:45.329013 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6f5d5cf467-m4grv" Dec 03 08:53:45 crc kubenswrapper[4573]: I1203 08:53:45.397673 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-dpwlm"] Dec 03 08:53:46 crc kubenswrapper[4573]: I1203 08:53:46.847301 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:46 crc kubenswrapper[4573]: I1203 08:53:46.847935 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:46 crc kubenswrapper[4573]: I1203 08:53:46.897388 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:47 crc kubenswrapper[4573]: I1203 08:53:47.345000 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" event={"ID":"483c74ea-46d0-42e4-84c3-35ed45a0507e","Type":"ContainerStarted","Data":"5fd9837732546c8d82e6149269b15b36f5f7783d6ba28fb1cb8880becb272e79"} Dec 03 08:53:47 crc kubenswrapper[4573]: I1203 08:53:47.372622 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-g2mdj" podStartSLOduration=3.648703109 podStartE2EDuration="14.372600204s" podCreationTimestamp="2025-12-03 08:53:33 +0000 UTC" firstStartedPulling="2025-12-03 08:53:35.570198591 +0000 UTC m=+936.138577850" lastFinishedPulling="2025-12-03 08:53:46.294095686 +0000 UTC m=+946.862474945" observedRunningTime="2025-12-03 08:53:47.370770244 +0000 UTC m=+947.939149533" watchObservedRunningTime="2025-12-03 08:53:47.372600204 +0000 UTC m=+947.940979463" Dec 03 08:53:47 crc kubenswrapper[4573]: I1203 08:53:47.399145 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:47 crc kubenswrapper[4573]: I1203 08:53:47.467992 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lc8zc"] Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.357093 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lc8zc" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerName="registry-server" containerID="cri-o://b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe" gracePeriod=2 Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.427389 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6mwnn" Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.766862 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.931670 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-catalog-content\") pod \"50c256a6-8345-4a1b-96f7-8475dcc28c47\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.931787 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6x8v\" (UniqueName: \"kubernetes.io/projected/50c256a6-8345-4a1b-96f7-8475dcc28c47-kube-api-access-x6x8v\") pod \"50c256a6-8345-4a1b-96f7-8475dcc28c47\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.931845 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-utilities\") pod \"50c256a6-8345-4a1b-96f7-8475dcc28c47\" (UID: \"50c256a6-8345-4a1b-96f7-8475dcc28c47\") " Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.933302 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-utilities" (OuterVolumeSpecName: "utilities") pod "50c256a6-8345-4a1b-96f7-8475dcc28c47" (UID: "50c256a6-8345-4a1b-96f7-8475dcc28c47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.942535 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c256a6-8345-4a1b-96f7-8475dcc28c47-kube-api-access-x6x8v" (OuterVolumeSpecName: "kube-api-access-x6x8v") pod "50c256a6-8345-4a1b-96f7-8475dcc28c47" (UID: "50c256a6-8345-4a1b-96f7-8475dcc28c47"). InnerVolumeSpecName "kube-api-access-x6x8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:53:49 crc kubenswrapper[4573]: I1203 08:53:49.985570 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50c256a6-8345-4a1b-96f7-8475dcc28c47" (UID: "50c256a6-8345-4a1b-96f7-8475dcc28c47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.034090 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.034639 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6x8v\" (UniqueName: \"kubernetes.io/projected/50c256a6-8345-4a1b-96f7-8475dcc28c47-kube-api-access-x6x8v\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.034759 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c256a6-8345-4a1b-96f7-8475dcc28c47-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.364403 4573 generic.go:334] "Generic (PLEG): container finished" podID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerID="b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe" exitCode=0 Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.364460 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc8zc" event={"ID":"50c256a6-8345-4a1b-96f7-8475dcc28c47","Type":"ContainerDied","Data":"b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe"} Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.364498 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lc8zc" event={"ID":"50c256a6-8345-4a1b-96f7-8475dcc28c47","Type":"ContainerDied","Data":"3de93eb9cc743fd43780bb53d327ff01fd9b409058076e742551b6a17192fc45"} Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.364519 4573 scope.go:117] "RemoveContainer" containerID="b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.364662 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lc8zc" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.392775 4573 scope.go:117] "RemoveContainer" containerID="024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.393731 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lc8zc"] Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.398389 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lc8zc"] Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.413417 4573 scope.go:117] "RemoveContainer" containerID="d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.433530 4573 scope.go:117] "RemoveContainer" containerID="b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe" Dec 03 08:53:50 crc kubenswrapper[4573]: E1203 08:53:50.434194 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe\": container with ID starting with b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe not found: ID does not exist" containerID="b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.434233 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe"} err="failed to get container status \"b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe\": rpc error: code = NotFound desc = could not find container \"b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe\": container with ID starting with b603713a5422045d14759c2606717ef5768197bd58f3276a2fb8efd46cc175fe not found: ID does not exist" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.434265 4573 scope.go:117] "RemoveContainer" containerID="024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986" Dec 03 08:53:50 crc kubenswrapper[4573]: E1203 08:53:50.434563 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986\": container with ID starting with 024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986 not found: ID does not exist" containerID="024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.434586 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986"} err="failed to get container status \"024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986\": rpc error: code = NotFound desc = could not find container \"024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986\": container with ID starting with 024189b5f64e6d3cd80f1236e417a4dce126ffdb7abfe664710ead3ac3aa8986 not found: ID does not exist" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.434600 4573 scope.go:117] "RemoveContainer" containerID="d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8" Dec 03 08:53:50 crc kubenswrapper[4573]: E1203 08:53:50.434840 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8\": container with ID starting with d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8 not found: ID does not exist" containerID="d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8" Dec 03 08:53:50 crc kubenswrapper[4573]: I1203 08:53:50.434869 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8"} err="failed to get container status \"d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8\": rpc error: code = NotFound desc = could not find container \"d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8\": container with ID starting with d59c968a14778f23b038b8f8d7a0529a876c20813b410ac7c4f92b0c15abd0b8 not found: ID does not exist" Dec 03 08:53:52 crc kubenswrapper[4573]: I1203 08:53:52.039636 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" path="/var/lib/kubelet/pods/50c256a6-8345-4a1b-96f7-8475dcc28c47/volumes" Dec 03 08:53:54 crc kubenswrapper[4573]: I1203 08:53:54.687503 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hlbnj" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.820617 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs"] Dec 03 08:54:08 crc kubenswrapper[4573]: E1203 08:54:08.823343 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerName="extract-utilities" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.823452 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerName="extract-utilities" Dec 03 08:54:08 crc kubenswrapper[4573]: E1203 08:54:08.823526 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerName="registry-server" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.823584 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerName="registry-server" Dec 03 08:54:08 crc kubenswrapper[4573]: E1203 08:54:08.823658 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerName="extract-content" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.823719 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerName="extract-content" Dec 03 08:54:08 crc kubenswrapper[4573]: E1203 08:54:08.823777 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="extract-content" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.823838 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="extract-content" Dec 03 08:54:08 crc kubenswrapper[4573]: E1203 08:54:08.824020 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="registry-server" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.824113 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="registry-server" Dec 03 08:54:08 crc kubenswrapper[4573]: E1203 08:54:08.824185 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="extract-utilities" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.824241 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="extract-utilities" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.824411 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ec8d9d8-d19c-4e2f-ba1d-31aaf7839966" containerName="registry-server" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.824475 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c256a6-8345-4a1b-96f7-8475dcc28c47" containerName="registry-server" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.826621 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.830374 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.839912 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs"] Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.882227 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.882350 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.882538 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqndn\" (UniqueName: \"kubernetes.io/projected/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-kube-api-access-pqndn\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.985040 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.985157 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqndn\" (UniqueName: \"kubernetes.io/projected/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-kube-api-access-pqndn\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.985195 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.985741 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:08 crc kubenswrapper[4573]: I1203 08:54:08.985745 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:09 crc kubenswrapper[4573]: I1203 08:54:09.008946 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqndn\" (UniqueName: \"kubernetes.io/projected/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-kube-api-access-pqndn\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:09 crc kubenswrapper[4573]: I1203 08:54:09.150083 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:09 crc kubenswrapper[4573]: I1203 08:54:09.645818 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs"] Dec 03 08:54:10 crc kubenswrapper[4573]: I1203 08:54:10.480036 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-dpwlm" podUID="bdc5518a-70d7-4f7f-bb7a-3320cb61f607" containerName="console" containerID="cri-o://c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef" gracePeriod=15 Dec 03 08:54:10 crc kubenswrapper[4573]: I1203 08:54:10.556171 4573 generic.go:334] "Generic (PLEG): container finished" podID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerID="0aeaa7c4623c620b624363928b00ac068262cb55c02b6849d8f8f2080537c8d2" exitCode=0 Dec 03 08:54:10 crc kubenswrapper[4573]: I1203 08:54:10.556245 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" event={"ID":"2860a6bf-c83f-4c6d-8d5a-062578bc40aa","Type":"ContainerDied","Data":"0aeaa7c4623c620b624363928b00ac068262cb55c02b6849d8f8f2080537c8d2"} Dec 03 08:54:10 crc kubenswrapper[4573]: I1203 08:54:10.556313 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" event={"ID":"2860a6bf-c83f-4c6d-8d5a-062578bc40aa","Type":"ContainerStarted","Data":"c312d3c250363e0453551ed77c2956d9c2e037880b5a046dcda2ea458c7d15a3"} Dec 03 08:54:10 crc kubenswrapper[4573]: I1203 08:54:10.882752 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-dpwlm_bdc5518a-70d7-4f7f-bb7a-3320cb61f607/console/0.log" Dec 03 08:54:10 crc kubenswrapper[4573]: I1203 08:54:10.883197 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.016789 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-config\") pod \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.016883 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-oauth-config\") pod \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.016955 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bpl4\" (UniqueName: \"kubernetes.io/projected/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-kube-api-access-2bpl4\") pod \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.017104 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-oauth-serving-cert\") pod \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.017229 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-serving-cert\") pod \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.017279 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-service-ca\") pod \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.017351 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-trusted-ca-bundle\") pod \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\" (UID: \"bdc5518a-70d7-4f7f-bb7a-3320cb61f607\") " Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.018360 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-config" (OuterVolumeSpecName: "console-config") pod "bdc5518a-70d7-4f7f-bb7a-3320cb61f607" (UID: "bdc5518a-70d7-4f7f-bb7a-3320cb61f607"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.018392 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "bdc5518a-70d7-4f7f-bb7a-3320cb61f607" (UID: "bdc5518a-70d7-4f7f-bb7a-3320cb61f607"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.018409 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "bdc5518a-70d7-4f7f-bb7a-3320cb61f607" (UID: "bdc5518a-70d7-4f7f-bb7a-3320cb61f607"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.019300 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-service-ca" (OuterVolumeSpecName: "service-ca") pod "bdc5518a-70d7-4f7f-bb7a-3320cb61f607" (UID: "bdc5518a-70d7-4f7f-bb7a-3320cb61f607"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.025979 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "bdc5518a-70d7-4f7f-bb7a-3320cb61f607" (UID: "bdc5518a-70d7-4f7f-bb7a-3320cb61f607"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.026447 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "bdc5518a-70d7-4f7f-bb7a-3320cb61f607" (UID: "bdc5518a-70d7-4f7f-bb7a-3320cb61f607"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.026959 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-kube-api-access-2bpl4" (OuterVolumeSpecName: "kube-api-access-2bpl4") pod "bdc5518a-70d7-4f7f-bb7a-3320cb61f607" (UID: "bdc5518a-70d7-4f7f-bb7a-3320cb61f607"). InnerVolumeSpecName "kube-api-access-2bpl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.118831 4573 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.118896 4573 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.118913 4573 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.118926 4573 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.118939 4573 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.118952 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bpl4\" (UniqueName: \"kubernetes.io/projected/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-kube-api-access-2bpl4\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.118968 4573 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bdc5518a-70d7-4f7f-bb7a-3320cb61f607-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.565644 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-dpwlm_bdc5518a-70d7-4f7f-bb7a-3320cb61f607/console/0.log" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.565715 4573 generic.go:334] "Generic (PLEG): container finished" podID="bdc5518a-70d7-4f7f-bb7a-3320cb61f607" containerID="c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef" exitCode=2 Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.565771 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dpwlm" event={"ID":"bdc5518a-70d7-4f7f-bb7a-3320cb61f607","Type":"ContainerDied","Data":"c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef"} Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.565830 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-dpwlm" event={"ID":"bdc5518a-70d7-4f7f-bb7a-3320cb61f607","Type":"ContainerDied","Data":"a450a21c80f33dcd780dd699f6b999e90b55e9c97d97b5019e1a10cd0b153ab0"} Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.565888 4573 scope.go:117] "RemoveContainer" containerID="c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.566202 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-dpwlm" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.589802 4573 scope.go:117] "RemoveContainer" containerID="c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef" Dec 03 08:54:11 crc kubenswrapper[4573]: E1203 08:54:11.590467 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef\": container with ID starting with c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef not found: ID does not exist" containerID="c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.590512 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef"} err="failed to get container status \"c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef\": rpc error: code = NotFound desc = could not find container \"c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef\": container with ID starting with c3afe3599e2bd3fd1c1a4aea96fafbce3c469bc25bde7b154e8a25cae27b28ef not found: ID does not exist" Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.622459 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-dpwlm"] Dec 03 08:54:11 crc kubenswrapper[4573]: I1203 08:54:11.629620 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-dpwlm"] Dec 03 08:54:12 crc kubenswrapper[4573]: I1203 08:54:12.040928 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdc5518a-70d7-4f7f-bb7a-3320cb61f607" path="/var/lib/kubelet/pods/bdc5518a-70d7-4f7f-bb7a-3320cb61f607/volumes" Dec 03 08:54:13 crc kubenswrapper[4573]: I1203 08:54:13.583816 4573 generic.go:334] "Generic (PLEG): container finished" podID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerID="fa72f1496f92ab854a48e1ea8bc67e80976128e2b39978002842b5fa2dc49dfb" exitCode=0 Dec 03 08:54:13 crc kubenswrapper[4573]: I1203 08:54:13.583886 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" event={"ID":"2860a6bf-c83f-4c6d-8d5a-062578bc40aa","Type":"ContainerDied","Data":"fa72f1496f92ab854a48e1ea8bc67e80976128e2b39978002842b5fa2dc49dfb"} Dec 03 08:54:14 crc kubenswrapper[4573]: I1203 08:54:14.594079 4573 generic.go:334] "Generic (PLEG): container finished" podID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerID="f9614a38e721d93c9806b31d9d260b345d28bca7ab5df778f3844d21f27b3963" exitCode=0 Dec 03 08:54:14 crc kubenswrapper[4573]: I1203 08:54:14.594596 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" event={"ID":"2860a6bf-c83f-4c6d-8d5a-062578bc40aa","Type":"ContainerDied","Data":"f9614a38e721d93c9806b31d9d260b345d28bca7ab5df778f3844d21f27b3963"} Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.180271 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.315313 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-bundle\") pod \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.315437 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-util\") pod \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.315545 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqndn\" (UniqueName: \"kubernetes.io/projected/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-kube-api-access-pqndn\") pod \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\" (UID: \"2860a6bf-c83f-4c6d-8d5a-062578bc40aa\") " Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.317571 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-bundle" (OuterVolumeSpecName: "bundle") pod "2860a6bf-c83f-4c6d-8d5a-062578bc40aa" (UID: "2860a6bf-c83f-4c6d-8d5a-062578bc40aa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.324315 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-kube-api-access-pqndn" (OuterVolumeSpecName: "kube-api-access-pqndn") pod "2860a6bf-c83f-4c6d-8d5a-062578bc40aa" (UID: "2860a6bf-c83f-4c6d-8d5a-062578bc40aa"). InnerVolumeSpecName "kube-api-access-pqndn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.339662 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-util" (OuterVolumeSpecName: "util") pod "2860a6bf-c83f-4c6d-8d5a-062578bc40aa" (UID: "2860a6bf-c83f-4c6d-8d5a-062578bc40aa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.418160 4573 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-util\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.418539 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqndn\" (UniqueName: \"kubernetes.io/projected/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-kube-api-access-pqndn\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.418610 4573 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2860a6bf-c83f-4c6d-8d5a-062578bc40aa-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.620329 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" event={"ID":"2860a6bf-c83f-4c6d-8d5a-062578bc40aa","Type":"ContainerDied","Data":"c312d3c250363e0453551ed77c2956d9c2e037880b5a046dcda2ea458c7d15a3"} Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.620416 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c312d3c250363e0453551ed77c2956d9c2e037880b5a046dcda2ea458c7d15a3" Dec 03 08:54:16 crc kubenswrapper[4573]: I1203 08:54:16.620551 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.943704 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.944759 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.993794 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6"] Dec 03 08:54:26 crc kubenswrapper[4573]: E1203 08:54:26.994199 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerName="extract" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.994220 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerName="extract" Dec 03 08:54:26 crc kubenswrapper[4573]: E1203 08:54:26.994239 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc5518a-70d7-4f7f-bb7a-3320cb61f607" containerName="console" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.994247 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc5518a-70d7-4f7f-bb7a-3320cb61f607" containerName="console" Dec 03 08:54:26 crc kubenswrapper[4573]: E1203 08:54:26.994259 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerName="util" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.994266 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerName="util" Dec 03 08:54:26 crc kubenswrapper[4573]: E1203 08:54:26.994281 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerName="pull" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.994287 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerName="pull" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.994397 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdc5518a-70d7-4f7f-bb7a-3320cb61f607" containerName="console" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.994409 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="2860a6bf-c83f-4c6d-8d5a-062578bc40aa" containerName="extract" Dec 03 08:54:26 crc kubenswrapper[4573]: I1203 08:54:26.994916 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.000631 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.000872 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.001009 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.001536 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-56hhs" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.001821 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.027769 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6"] Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.099563 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29c77033-f8ba-44b5-a275-0861c60318bf-apiservice-cert\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.100287 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttht2\" (UniqueName: \"kubernetes.io/projected/29c77033-f8ba-44b5-a275-0861c60318bf-kube-api-access-ttht2\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.100626 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29c77033-f8ba-44b5-a275-0861c60318bf-webhook-cert\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.202597 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttht2\" (UniqueName: \"kubernetes.io/projected/29c77033-f8ba-44b5-a275-0861c60318bf-kube-api-access-ttht2\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.202722 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29c77033-f8ba-44b5-a275-0861c60318bf-webhook-cert\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.202837 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29c77033-f8ba-44b5-a275-0861c60318bf-apiservice-cert\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.210557 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29c77033-f8ba-44b5-a275-0861c60318bf-webhook-cert\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.215915 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29c77033-f8ba-44b5-a275-0861c60318bf-apiservice-cert\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.231065 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttht2\" (UniqueName: \"kubernetes.io/projected/29c77033-f8ba-44b5-a275-0861c60318bf-kube-api-access-ttht2\") pod \"metallb-operator-controller-manager-d75bb676c-wwfz6\" (UID: \"29c77033-f8ba-44b5-a275-0861c60318bf\") " pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.313362 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.684903 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh"] Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.688488 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.703127 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.703700 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.704297 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pszt9" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.712443 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e051555-a35f-45f0-ac65-237da26aa296-apiservice-cert\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.712516 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e051555-a35f-45f0-ac65-237da26aa296-webhook-cert\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.712602 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49qwr\" (UniqueName: \"kubernetes.io/projected/4e051555-a35f-45f0-ac65-237da26aa296-kube-api-access-49qwr\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.732151 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh"] Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.814095 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49qwr\" (UniqueName: \"kubernetes.io/projected/4e051555-a35f-45f0-ac65-237da26aa296-kube-api-access-49qwr\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.814200 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e051555-a35f-45f0-ac65-237da26aa296-apiservice-cert\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.814252 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e051555-a35f-45f0-ac65-237da26aa296-webhook-cert\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.822180 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4e051555-a35f-45f0-ac65-237da26aa296-apiservice-cert\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.845627 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4e051555-a35f-45f0-ac65-237da26aa296-webhook-cert\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.851038 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49qwr\" (UniqueName: \"kubernetes.io/projected/4e051555-a35f-45f0-ac65-237da26aa296-kube-api-access-49qwr\") pod \"metallb-operator-webhook-server-b9cfd44fd-mswlh\" (UID: \"4e051555-a35f-45f0-ac65-237da26aa296\") " pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:27 crc kubenswrapper[4573]: I1203 08:54:27.865568 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6"] Dec 03 08:54:27 crc kubenswrapper[4573]: W1203 08:54:27.892530 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29c77033_f8ba_44b5_a275_0861c60318bf.slice/crio-1201578c63691720263810a1a17708882683264f1b49c3fed68c82b59248f140 WatchSource:0}: Error finding container 1201578c63691720263810a1a17708882683264f1b49c3fed68c82b59248f140: Status 404 returned error can't find the container with id 1201578c63691720263810a1a17708882683264f1b49c3fed68c82b59248f140 Dec 03 08:54:28 crc kubenswrapper[4573]: I1203 08:54:28.027889 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:28 crc kubenswrapper[4573]: I1203 08:54:28.501855 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh"] Dec 03 08:54:28 crc kubenswrapper[4573]: W1203 08:54:28.519317 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e051555_a35f_45f0_ac65_237da26aa296.slice/crio-3026e5dd56ff7ccdf5bc12a21b0f61c4cdc93e395834f0bbb78877b4de457559 WatchSource:0}: Error finding container 3026e5dd56ff7ccdf5bc12a21b0f61c4cdc93e395834f0bbb78877b4de457559: Status 404 returned error can't find the container with id 3026e5dd56ff7ccdf5bc12a21b0f61c4cdc93e395834f0bbb78877b4de457559 Dec 03 08:54:28 crc kubenswrapper[4573]: I1203 08:54:28.716389 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" event={"ID":"4e051555-a35f-45f0-ac65-237da26aa296","Type":"ContainerStarted","Data":"3026e5dd56ff7ccdf5bc12a21b0f61c4cdc93e395834f0bbb78877b4de457559"} Dec 03 08:54:28 crc kubenswrapper[4573]: I1203 08:54:28.717885 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" event={"ID":"29c77033-f8ba-44b5-a275-0861c60318bf","Type":"ContainerStarted","Data":"1201578c63691720263810a1a17708882683264f1b49c3fed68c82b59248f140"} Dec 03 08:54:36 crc kubenswrapper[4573]: I1203 08:54:36.784107 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" event={"ID":"4e051555-a35f-45f0-ac65-237da26aa296","Type":"ContainerStarted","Data":"35b478fce18df5f0f148b2712338a96662cee649d91c2650126dcddc5c531724"} Dec 03 08:54:36 crc kubenswrapper[4573]: I1203 08:54:36.785461 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:36 crc kubenswrapper[4573]: I1203 08:54:36.787424 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" event={"ID":"29c77033-f8ba-44b5-a275-0861c60318bf","Type":"ContainerStarted","Data":"b0113145f44974e96909fc19918695576098586e1003b6e97d811dd0605e6ee6"} Dec 03 08:54:36 crc kubenswrapper[4573]: I1203 08:54:36.787588 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:54:36 crc kubenswrapper[4573]: I1203 08:54:36.818549 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" podStartSLOduration=2.527984447 podStartE2EDuration="9.818524408s" podCreationTimestamp="2025-12-03 08:54:27 +0000 UTC" firstStartedPulling="2025-12-03 08:54:28.522885758 +0000 UTC m=+989.091265027" lastFinishedPulling="2025-12-03 08:54:35.813425729 +0000 UTC m=+996.381804988" observedRunningTime="2025-12-03 08:54:36.816145683 +0000 UTC m=+997.384524942" watchObservedRunningTime="2025-12-03 08:54:36.818524408 +0000 UTC m=+997.386903667" Dec 03 08:54:36 crc kubenswrapper[4573]: I1203 08:54:36.852418 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" podStartSLOduration=2.956221545 podStartE2EDuration="10.85239362s" podCreationTimestamp="2025-12-03 08:54:26 +0000 UTC" firstStartedPulling="2025-12-03 08:54:27.896444237 +0000 UTC m=+988.464823506" lastFinishedPulling="2025-12-03 08:54:35.792616322 +0000 UTC m=+996.360995581" observedRunningTime="2025-12-03 08:54:36.851423373 +0000 UTC m=+997.419802632" watchObservedRunningTime="2025-12-03 08:54:36.85239362 +0000 UTC m=+997.420772879" Dec 03 08:54:48 crc kubenswrapper[4573]: I1203 08:54:48.038124 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-b9cfd44fd-mswlh" Dec 03 08:54:56 crc kubenswrapper[4573]: I1203 08:54:56.943863 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:54:56 crc kubenswrapper[4573]: I1203 08:54:56.944851 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:55:07 crc kubenswrapper[4573]: I1203 08:55:07.316553 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-d75bb676c-wwfz6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.248683 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6"] Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.249671 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.254515 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.255418 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6fjjm"] Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.256546 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-z2gd8" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.258794 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.263390 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.264135 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277230 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-conf\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277302 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e088580-c78b-42cc-b013-bc58bcffa838-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-cngn6\" (UID: \"0e088580-c78b-42cc-b013-bc58bcffa838\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277352 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-sockets\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277374 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9msk\" (UniqueName: \"kubernetes.io/projected/0e088580-c78b-42cc-b013-bc58bcffa838-kube-api-access-g9msk\") pod \"frr-k8s-webhook-server-7fcb986d4-cngn6\" (UID: \"0e088580-c78b-42cc-b013-bc58bcffa838\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277397 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-startup\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277420 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw4xs\" (UniqueName: \"kubernetes.io/projected/2d98f6ef-9108-408a-8ae6-b1116c434d68-kube-api-access-jw4xs\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277468 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-reloader\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277490 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics-certs\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.277534 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.320663 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6"] Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379029 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-reloader\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379098 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics-certs\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379424 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379455 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-conf\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379481 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e088580-c78b-42cc-b013-bc58bcffa838-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-cngn6\" (UID: \"0e088580-c78b-42cc-b013-bc58bcffa838\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379535 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-sockets\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379566 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9msk\" (UniqueName: \"kubernetes.io/projected/0e088580-c78b-42cc-b013-bc58bcffa838-kube-api-access-g9msk\") pod \"frr-k8s-webhook-server-7fcb986d4-cngn6\" (UID: \"0e088580-c78b-42cc-b013-bc58bcffa838\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379586 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-startup\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379610 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw4xs\" (UniqueName: \"kubernetes.io/projected/2d98f6ef-9108-408a-8ae6-b1116c434d68-kube-api-access-jw4xs\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.379656 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-reloader\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: E1203 08:55:08.380294 4573 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 03 08:55:08 crc kubenswrapper[4573]: E1203 08:55:08.380385 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics-certs podName:2d98f6ef-9108-408a-8ae6-b1116c434d68 nodeName:}" failed. No retries permitted until 2025-12-03 08:55:08.88036047 +0000 UTC m=+1029.448739909 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics-certs") pod "frr-k8s-6fjjm" (UID: "2d98f6ef-9108-408a-8ae6-b1116c434d68") : secret "frr-k8s-certs-secret" not found Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.380297 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-sockets\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.380825 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-conf\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.380863 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.381332 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2d98f6ef-9108-408a-8ae6-b1116c434d68-frr-startup\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.389171 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e088580-c78b-42cc-b013-bc58bcffa838-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-cngn6\" (UID: \"0e088580-c78b-42cc-b013-bc58bcffa838\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.406720 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9msk\" (UniqueName: \"kubernetes.io/projected/0e088580-c78b-42cc-b013-bc58bcffa838-kube-api-access-g9msk\") pod \"frr-k8s-webhook-server-7fcb986d4-cngn6\" (UID: \"0e088580-c78b-42cc-b013-bc58bcffa838\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.428738 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw4xs\" (UniqueName: \"kubernetes.io/projected/2d98f6ef-9108-408a-8ae6-b1116c434d68-kube-api-access-jw4xs\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.429417 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hw9mt"] Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.436713 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.440538 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.440847 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.441340 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-nbsvb" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.441530 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.480665 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/09e37261-797a-4fb1-9ae2-5de662b865e8-metallb-excludel2\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.480726 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqrzj\" (UniqueName: \"kubernetes.io/projected/09e37261-797a-4fb1-9ae2-5de662b865e8-kube-api-access-jqrzj\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.480765 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.480817 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-metrics-certs\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.481591 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-rj7p6"] Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.482738 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.486191 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.508680 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-rj7p6"] Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.584169 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-cert\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.584281 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/09e37261-797a-4fb1-9ae2-5de662b865e8-metallb-excludel2\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.584338 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqrzj\" (UniqueName: \"kubernetes.io/projected/09e37261-797a-4fb1-9ae2-5de662b865e8-kube-api-access-jqrzj\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.584373 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.584398 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sngv8\" (UniqueName: \"kubernetes.io/projected/974ec28d-5851-4dd3-b992-9920ed8d32de-kube-api-access-sngv8\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.584458 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-metrics-certs\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.584477 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-metrics-certs\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: E1203 08:55:08.584936 4573 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 08:55:08 crc kubenswrapper[4573]: E1203 08:55:08.585114 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist podName:09e37261-797a-4fb1-9ae2-5de662b865e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:55:09.085090983 +0000 UTC m=+1029.653470242 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist") pod "speaker-hw9mt" (UID: "09e37261-797a-4fb1-9ae2-5de662b865e8") : secret "metallb-memberlist" not found Dec 03 08:55:08 crc kubenswrapper[4573]: E1203 08:55:08.584992 4573 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 03 08:55:08 crc kubenswrapper[4573]: E1203 08:55:08.585327 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-metrics-certs podName:09e37261-797a-4fb1-9ae2-5de662b865e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:55:09.085314459 +0000 UTC m=+1029.653693718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-metrics-certs") pod "speaker-hw9mt" (UID: "09e37261-797a-4fb1-9ae2-5de662b865e8") : secret "speaker-certs-secret" not found Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.585563 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/09e37261-797a-4fb1-9ae2-5de662b865e8-metallb-excludel2\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.585791 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.611351 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqrzj\" (UniqueName: \"kubernetes.io/projected/09e37261-797a-4fb1-9ae2-5de662b865e8-kube-api-access-jqrzj\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.685455 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-cert\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.685937 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sngv8\" (UniqueName: \"kubernetes.io/projected/974ec28d-5851-4dd3-b992-9920ed8d32de-kube-api-access-sngv8\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.686007 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-metrics-certs\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: E1203 08:55:08.686114 4573 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 03 08:55:08 crc kubenswrapper[4573]: E1203 08:55:08.686191 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-metrics-certs podName:974ec28d-5851-4dd3-b992-9920ed8d32de nodeName:}" failed. No retries permitted until 2025-12-03 08:55:09.186170324 +0000 UTC m=+1029.754549583 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-metrics-certs") pod "controller-f8648f98b-rj7p6" (UID: "974ec28d-5851-4dd3-b992-9920ed8d32de") : secret "controller-certs-secret" not found Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.689068 4573 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.701396 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-cert\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.708294 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sngv8\" (UniqueName: \"kubernetes.io/projected/974ec28d-5851-4dd3-b992-9920ed8d32de-kube-api-access-sngv8\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.890379 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics-certs\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.894436 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2d98f6ef-9108-408a-8ae6-b1116c434d68-metrics-certs\") pod \"frr-k8s-6fjjm\" (UID: \"2d98f6ef-9108-408a-8ae6-b1116c434d68\") " pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.897335 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:08 crc kubenswrapper[4573]: I1203 08:55:08.975935 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6"] Dec 03 08:55:09 crc kubenswrapper[4573]: I1203 08:55:09.016698 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" event={"ID":"0e088580-c78b-42cc-b013-bc58bcffa838","Type":"ContainerStarted","Data":"5dad27e785007f81866bf4616cba598743f5b238af99537272c9e8b7a18f3a95"} Dec 03 08:55:09 crc kubenswrapper[4573]: I1203 08:55:09.093107 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-metrics-certs\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:09 crc kubenswrapper[4573]: I1203 08:55:09.093300 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:09 crc kubenswrapper[4573]: E1203 08:55:09.093579 4573 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 08:55:09 crc kubenswrapper[4573]: E1203 08:55:09.093746 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist podName:09e37261-797a-4fb1-9ae2-5de662b865e8 nodeName:}" failed. No retries permitted until 2025-12-03 08:55:10.093709988 +0000 UTC m=+1030.662089417 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist") pod "speaker-hw9mt" (UID: "09e37261-797a-4fb1-9ae2-5de662b865e8") : secret "metallb-memberlist" not found Dec 03 08:55:09 crc kubenswrapper[4573]: I1203 08:55:09.099718 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-metrics-certs\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:09 crc kubenswrapper[4573]: I1203 08:55:09.194917 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-metrics-certs\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:09 crc kubenswrapper[4573]: I1203 08:55:09.199668 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/974ec28d-5851-4dd3-b992-9920ed8d32de-metrics-certs\") pod \"controller-f8648f98b-rj7p6\" (UID: \"974ec28d-5851-4dd3-b992-9920ed8d32de\") " pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:09 crc kubenswrapper[4573]: I1203 08:55:09.424938 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:09 crc kubenswrapper[4573]: I1203 08:55:09.695766 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-rj7p6"] Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.025283 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerStarted","Data":"e082317df4dd8587676cebf52ef1c659ffec3aec65c1152b01c600081e1025ae"} Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.027631 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rj7p6" event={"ID":"974ec28d-5851-4dd3-b992-9920ed8d32de","Type":"ContainerStarted","Data":"68eec4c71f145cdeee77abe027004efe02523d6325aa3ea7bb69b8560a6abafb"} Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.027659 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rj7p6" event={"ID":"974ec28d-5851-4dd3-b992-9920ed8d32de","Type":"ContainerStarted","Data":"307c2254c783fadf96aaaa285c5de5ab806f8508e79e9317d30d9825f28fc590"} Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.027671 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-rj7p6" event={"ID":"974ec28d-5851-4dd3-b992-9920ed8d32de","Type":"ContainerStarted","Data":"bdf30c13190f2e7406671c1e49256b93dd183ecc108e08dcf8f14439cb23b6b7"} Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.027892 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.119382 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.140037 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/09e37261-797a-4fb1-9ae2-5de662b865e8-memberlist\") pod \"speaker-hw9mt\" (UID: \"09e37261-797a-4fb1-9ae2-5de662b865e8\") " pod="metallb-system/speaker-hw9mt" Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.140902 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-rj7p6" podStartSLOduration=2.140874892 podStartE2EDuration="2.140874892s" podCreationTimestamp="2025-12-03 08:55:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:55:10.140607645 +0000 UTC m=+1030.708986904" watchObservedRunningTime="2025-12-03 08:55:10.140874892 +0000 UTC m=+1030.709254151" Dec 03 08:55:10 crc kubenswrapper[4573]: I1203 08:55:10.315571 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hw9mt" Dec 03 08:55:10 crc kubenswrapper[4573]: W1203 08:55:10.350834 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09e37261_797a_4fb1_9ae2_5de662b865e8.slice/crio-196cc7107fe6caffe0dbb8fbdbc0a17e99d9f9009e5f8b008062aebeecff5026 WatchSource:0}: Error finding container 196cc7107fe6caffe0dbb8fbdbc0a17e99d9f9009e5f8b008062aebeecff5026: Status 404 returned error can't find the container with id 196cc7107fe6caffe0dbb8fbdbc0a17e99d9f9009e5f8b008062aebeecff5026 Dec 03 08:55:11 crc kubenswrapper[4573]: I1203 08:55:11.210859 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hw9mt" event={"ID":"09e37261-797a-4fb1-9ae2-5de662b865e8","Type":"ContainerStarted","Data":"8d7ee7b3322870baf6a4f0ec616e42ced2fa6bbbe3e2a5d32a4651c89a283ecf"} Dec 03 08:55:11 crc kubenswrapper[4573]: I1203 08:55:11.210966 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hw9mt" event={"ID":"09e37261-797a-4fb1-9ae2-5de662b865e8","Type":"ContainerStarted","Data":"196cc7107fe6caffe0dbb8fbdbc0a17e99d9f9009e5f8b008062aebeecff5026"} Dec 03 08:55:12 crc kubenswrapper[4573]: I1203 08:55:12.235670 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hw9mt" event={"ID":"09e37261-797a-4fb1-9ae2-5de662b865e8","Type":"ContainerStarted","Data":"5d12ff4020762acb40580075764d0f3abbc95c33fa8d7c53a30c71551abecd66"} Dec 03 08:55:12 crc kubenswrapper[4573]: I1203 08:55:12.236216 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hw9mt" Dec 03 08:55:12 crc kubenswrapper[4573]: I1203 08:55:12.346214 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hw9mt" podStartSLOduration=4.346173361 podStartE2EDuration="4.346173361s" podCreationTimestamp="2025-12-03 08:55:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:55:12.344766593 +0000 UTC m=+1032.913145852" watchObservedRunningTime="2025-12-03 08:55:12.346173361 +0000 UTC m=+1032.914552610" Dec 03 08:55:19 crc kubenswrapper[4573]: I1203 08:55:19.432956 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-rj7p6" Dec 03 08:55:20 crc kubenswrapper[4573]: I1203 08:55:20.320934 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hw9mt" Dec 03 08:55:23 crc kubenswrapper[4573]: I1203 08:55:23.511273 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" event={"ID":"0e088580-c78b-42cc-b013-bc58bcffa838","Type":"ContainerStarted","Data":"ddfa9571fd7839da81b36e890dec01a42dff6d1b226119ceadf7ddcabe7f5a55"} Dec 03 08:55:23 crc kubenswrapper[4573]: I1203 08:55:23.512379 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:23 crc kubenswrapper[4573]: I1203 08:55:23.515654 4573 generic.go:334] "Generic (PLEG): container finished" podID="2d98f6ef-9108-408a-8ae6-b1116c434d68" containerID="0f2580c1f02d010d6cbce96275c5bf87da36cb651b1fff279165ce2b52b75b06" exitCode=0 Dec 03 08:55:23 crc kubenswrapper[4573]: I1203 08:55:23.515702 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerDied","Data":"0f2580c1f02d010d6cbce96275c5bf87da36cb651b1fff279165ce2b52b75b06"} Dec 03 08:55:23 crc kubenswrapper[4573]: I1203 08:55:23.547941 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" podStartSLOduration=1.4051227640000001 podStartE2EDuration="15.547911493s" podCreationTimestamp="2025-12-03 08:55:08 +0000 UTC" firstStartedPulling="2025-12-03 08:55:08.998159207 +0000 UTC m=+1029.566538466" lastFinishedPulling="2025-12-03 08:55:23.140947936 +0000 UTC m=+1043.709327195" observedRunningTime="2025-12-03 08:55:23.546423093 +0000 UTC m=+1044.114802372" watchObservedRunningTime="2025-12-03 08:55:23.547911493 +0000 UTC m=+1044.116290752" Dec 03 08:55:23 crc kubenswrapper[4573]: I1203 08:55:23.954993 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qgw4z"] Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.008649 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qgw4z" Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.014959 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.015185 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-p2hbn" Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.015638 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.018647 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qgw4z"] Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.145083 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdhgr\" (UniqueName: \"kubernetes.io/projected/44a24893-6416-449f-bc55-b09cb4636117-kube-api-access-fdhgr\") pod \"openstack-operator-index-qgw4z\" (UID: \"44a24893-6416-449f-bc55-b09cb4636117\") " pod="openstack-operators/openstack-operator-index-qgw4z" Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.246409 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdhgr\" (UniqueName: \"kubernetes.io/projected/44a24893-6416-449f-bc55-b09cb4636117-kube-api-access-fdhgr\") pod \"openstack-operator-index-qgw4z\" (UID: \"44a24893-6416-449f-bc55-b09cb4636117\") " pod="openstack-operators/openstack-operator-index-qgw4z" Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.270496 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdhgr\" (UniqueName: \"kubernetes.io/projected/44a24893-6416-449f-bc55-b09cb4636117-kube-api-access-fdhgr\") pod \"openstack-operator-index-qgw4z\" (UID: \"44a24893-6416-449f-bc55-b09cb4636117\") " pod="openstack-operators/openstack-operator-index-qgw4z" Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.349385 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qgw4z" Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.570515 4573 generic.go:334] "Generic (PLEG): container finished" podID="2d98f6ef-9108-408a-8ae6-b1116c434d68" containerID="b3ea2dcde30a78380b21add8c0599fc8a3652a298435eaa8d7f2ad51e2a84bf3" exitCode=0 Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.572011 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerDied","Data":"b3ea2dcde30a78380b21add8c0599fc8a3652a298435eaa8d7f2ad51e2a84bf3"} Dec 03 08:55:24 crc kubenswrapper[4573]: I1203 08:55:24.801311 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qgw4z"] Dec 03 08:55:24 crc kubenswrapper[4573]: W1203 08:55:24.809592 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44a24893_6416_449f_bc55_b09cb4636117.slice/crio-127587ba171548d85b87b254f7be60d1f0e778b0e18710954300f28a1615d045 WatchSource:0}: Error finding container 127587ba171548d85b87b254f7be60d1f0e778b0e18710954300f28a1615d045: Status 404 returned error can't find the container with id 127587ba171548d85b87b254f7be60d1f0e778b0e18710954300f28a1615d045 Dec 03 08:55:25 crc kubenswrapper[4573]: I1203 08:55:25.592737 4573 generic.go:334] "Generic (PLEG): container finished" podID="2d98f6ef-9108-408a-8ae6-b1116c434d68" containerID="8347e47477685307d87356c99abc0ffaee5cc062818ff3bd4c2a16d98db4f339" exitCode=0 Dec 03 08:55:25 crc kubenswrapper[4573]: I1203 08:55:25.592842 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerDied","Data":"8347e47477685307d87356c99abc0ffaee5cc062818ff3bd4c2a16d98db4f339"} Dec 03 08:55:25 crc kubenswrapper[4573]: I1203 08:55:25.597328 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qgw4z" event={"ID":"44a24893-6416-449f-bc55-b09cb4636117","Type":"ContainerStarted","Data":"127587ba171548d85b87b254f7be60d1f0e778b0e18710954300f28a1615d045"} Dec 03 08:55:26 crc kubenswrapper[4573]: I1203 08:55:26.943267 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:55:26 crc kubenswrapper[4573]: I1203 08:55:26.943815 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:55:26 crc kubenswrapper[4573]: I1203 08:55:26.944161 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:55:26 crc kubenswrapper[4573]: I1203 08:55:26.945237 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"68d9133084811f469a486ce4e9540b723fb96d8cd528443eac595b0a1c6329ed"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:55:26 crc kubenswrapper[4573]: I1203 08:55:26.945322 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://68d9133084811f469a486ce4e9540b723fb96d8cd528443eac595b0a1c6329ed" gracePeriod=600 Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.325116 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qgw4z"] Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.631323 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerStarted","Data":"6ec37795b08b9f729b4538cae7bcf67cfab17b21d6fed8f1d4f2b5d294b0adb9"} Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.638584 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="68d9133084811f469a486ce4e9540b723fb96d8cd528443eac595b0a1c6329ed" exitCode=0 Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.638646 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"68d9133084811f469a486ce4e9540b723fb96d8cd528443eac595b0a1c6329ed"} Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.638683 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"93ffc1f0cd99ae07290bc9e27b4a1d6c063abbb15b748bb21e1aa18b0c38a447"} Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.638702 4573 scope.go:117] "RemoveContainer" containerID="e638f4437cd42f13a1acdd69aa087a2d6a308b827f8a1fbde0eba4442f1d564e" Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.930671 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-h4tdp"] Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.931846 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:27 crc kubenswrapper[4573]: I1203 08:55:27.955238 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-h4tdp"] Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.036240 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpdnn\" (UniqueName: \"kubernetes.io/projected/1b63239f-176e-4a3b-ac7d-605be09bd269-kube-api-access-wpdnn\") pod \"openstack-operator-index-h4tdp\" (UID: \"1b63239f-176e-4a3b-ac7d-605be09bd269\") " pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.138452 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpdnn\" (UniqueName: \"kubernetes.io/projected/1b63239f-176e-4a3b-ac7d-605be09bd269-kube-api-access-wpdnn\") pod \"openstack-operator-index-h4tdp\" (UID: \"1b63239f-176e-4a3b-ac7d-605be09bd269\") " pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.176316 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpdnn\" (UniqueName: \"kubernetes.io/projected/1b63239f-176e-4a3b-ac7d-605be09bd269-kube-api-access-wpdnn\") pod \"openstack-operator-index-h4tdp\" (UID: \"1b63239f-176e-4a3b-ac7d-605be09bd269\") " pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.250366 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.660235 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerStarted","Data":"a5e3e7b568d467afe24029e8438a321d97466dbf3ede650f9dfa00e5e34ebd4f"} Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.661137 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerStarted","Data":"f98098385b09daabf39626979ce0ea6155fc134f9ad19cfe98065870726320a8"} Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.661155 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerStarted","Data":"9d826d8fa7c8ae4346d75bd7a56ed13ea1ecca78c2e5d713f97a80452d80b457"} Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.661186 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerStarted","Data":"18838c1deadd49ca146d702e747a83435f3e498f1dd1ffbb2052da450f824d39"} Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.669510 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qgw4z" event={"ID":"44a24893-6416-449f-bc55-b09cb4636117","Type":"ContainerStarted","Data":"6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098"} Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.669690 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-qgw4z" podUID="44a24893-6416-449f-bc55-b09cb4636117" containerName="registry-server" containerID="cri-o://6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098" gracePeriod=2 Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.694685 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qgw4z" podStartSLOduration=3.049232411 podStartE2EDuration="5.694656389s" podCreationTimestamp="2025-12-03 08:55:23 +0000 UTC" firstStartedPulling="2025-12-03 08:55:24.817677207 +0000 UTC m=+1045.386056466" lastFinishedPulling="2025-12-03 08:55:27.463101175 +0000 UTC m=+1048.031480444" observedRunningTime="2025-12-03 08:55:28.694481583 +0000 UTC m=+1049.262860842" watchObservedRunningTime="2025-12-03 08:55:28.694656389 +0000 UTC m=+1049.263035648" Dec 03 08:55:28 crc kubenswrapper[4573]: I1203 08:55:28.828538 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-h4tdp"] Dec 03 08:55:28 crc kubenswrapper[4573]: W1203 08:55:28.846585 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b63239f_176e_4a3b_ac7d_605be09bd269.slice/crio-f9cdeefa6721b4ee81f2407f24e4fd15fe3782c71362b8597c5204570fdb96eb WatchSource:0}: Error finding container f9cdeefa6721b4ee81f2407f24e4fd15fe3782c71362b8597c5204570fdb96eb: Status 404 returned error can't find the container with id f9cdeefa6721b4ee81f2407f24e4fd15fe3782c71362b8597c5204570fdb96eb Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.193821 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qgw4z" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.380813 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdhgr\" (UniqueName: \"kubernetes.io/projected/44a24893-6416-449f-bc55-b09cb4636117-kube-api-access-fdhgr\") pod \"44a24893-6416-449f-bc55-b09cb4636117\" (UID: \"44a24893-6416-449f-bc55-b09cb4636117\") " Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.388743 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44a24893-6416-449f-bc55-b09cb4636117-kube-api-access-fdhgr" (OuterVolumeSpecName: "kube-api-access-fdhgr") pod "44a24893-6416-449f-bc55-b09cb4636117" (UID: "44a24893-6416-449f-bc55-b09cb4636117"). InnerVolumeSpecName "kube-api-access-fdhgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.482994 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdhgr\" (UniqueName: \"kubernetes.io/projected/44a24893-6416-449f-bc55-b09cb4636117-kube-api-access-fdhgr\") on node \"crc\" DevicePath \"\"" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.684218 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6fjjm" event={"ID":"2d98f6ef-9108-408a-8ae6-b1116c434d68","Type":"ContainerStarted","Data":"ee644eaebf80c32755441eaca59e401d2d15bdac106d8d335469c0167dcc591d"} Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.684441 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.688405 4573 generic.go:334] "Generic (PLEG): container finished" podID="44a24893-6416-449f-bc55-b09cb4636117" containerID="6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098" exitCode=0 Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.688614 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qgw4z" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.689258 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qgw4z" event={"ID":"44a24893-6416-449f-bc55-b09cb4636117","Type":"ContainerDied","Data":"6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098"} Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.689495 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qgw4z" event={"ID":"44a24893-6416-449f-bc55-b09cb4636117","Type":"ContainerDied","Data":"127587ba171548d85b87b254f7be60d1f0e778b0e18710954300f28a1615d045"} Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.689537 4573 scope.go:117] "RemoveContainer" containerID="6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.693327 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-h4tdp" event={"ID":"1b63239f-176e-4a3b-ac7d-605be09bd269","Type":"ContainerStarted","Data":"99ef669681a3217e01519b8e5b0d90c0a2c6ebf397975e112ac01934694094a9"} Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.693371 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-h4tdp" event={"ID":"1b63239f-176e-4a3b-ac7d-605be09bd269","Type":"ContainerStarted","Data":"f9cdeefa6721b4ee81f2407f24e4fd15fe3782c71362b8597c5204570fdb96eb"} Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.726983 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6fjjm" podStartSLOduration=7.628387041 podStartE2EDuration="21.726956397s" podCreationTimestamp="2025-12-03 08:55:08 +0000 UTC" firstStartedPulling="2025-12-03 08:55:09.065039097 +0000 UTC m=+1029.633418356" lastFinishedPulling="2025-12-03 08:55:23.163608453 +0000 UTC m=+1043.731987712" observedRunningTime="2025-12-03 08:55:29.719627658 +0000 UTC m=+1050.288006907" watchObservedRunningTime="2025-12-03 08:55:29.726956397 +0000 UTC m=+1050.295335656" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.731343 4573 scope.go:117] "RemoveContainer" containerID="6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098" Dec 03 08:55:29 crc kubenswrapper[4573]: E1203 08:55:29.732105 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098\": container with ID starting with 6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098 not found: ID does not exist" containerID="6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.732145 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098"} err="failed to get container status \"6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098\": rpc error: code = NotFound desc = could not find container \"6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098\": container with ID starting with 6f1b25baa5b7ac2849ed88d4f0c1ba90fbac7c11aba17abd3febe889ba21c098 not found: ID does not exist" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.745346 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-h4tdp" podStartSLOduration=2.691995866 podStartE2EDuration="2.745317748s" podCreationTimestamp="2025-12-03 08:55:27 +0000 UTC" firstStartedPulling="2025-12-03 08:55:28.851690303 +0000 UTC m=+1049.420069562" lastFinishedPulling="2025-12-03 08:55:28.905012185 +0000 UTC m=+1049.473391444" observedRunningTime="2025-12-03 08:55:29.742498391 +0000 UTC m=+1050.310877680" watchObservedRunningTime="2025-12-03 08:55:29.745317748 +0000 UTC m=+1050.313697007" Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.762488 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qgw4z"] Dec 03 08:55:29 crc kubenswrapper[4573]: I1203 08:55:29.770260 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-qgw4z"] Dec 03 08:55:30 crc kubenswrapper[4573]: I1203 08:55:30.040963 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44a24893-6416-449f-bc55-b09cb4636117" path="/var/lib/kubelet/pods/44a24893-6416-449f-bc55-b09cb4636117/volumes" Dec 03 08:55:33 crc kubenswrapper[4573]: I1203 08:55:33.897960 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:33 crc kubenswrapper[4573]: I1203 08:55:33.948195 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:38 crc kubenswrapper[4573]: I1203 08:55:38.251602 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:38 crc kubenswrapper[4573]: I1203 08:55:38.254213 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:38 crc kubenswrapper[4573]: I1203 08:55:38.282000 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:38 crc kubenswrapper[4573]: I1203 08:55:38.591589 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cngn6" Dec 03 08:55:38 crc kubenswrapper[4573]: I1203 08:55:38.788881 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-h4tdp" Dec 03 08:55:38 crc kubenswrapper[4573]: I1203 08:55:38.901417 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6fjjm" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.376918 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp"] Dec 03 08:55:41 crc kubenswrapper[4573]: E1203 08:55:41.377296 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a24893-6416-449f-bc55-b09cb4636117" containerName="registry-server" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.377312 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a24893-6416-449f-bc55-b09cb4636117" containerName="registry-server" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.377490 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="44a24893-6416-449f-bc55-b09cb4636117" containerName="registry-server" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.378630 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.381662 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-kvs2t" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.396731 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp"] Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.569950 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-util\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.570106 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-bundle\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.570147 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgrmg\" (UniqueName: \"kubernetes.io/projected/c09528b8-a5a1-490d-86ef-57027d97f282-kube-api-access-xgrmg\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.671600 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-util\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.671682 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-bundle\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.671716 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgrmg\" (UniqueName: \"kubernetes.io/projected/c09528b8-a5a1-490d-86ef-57027d97f282-kube-api-access-xgrmg\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.672684 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-bundle\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.672840 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-util\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.695004 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgrmg\" (UniqueName: \"kubernetes.io/projected/c09528b8-a5a1-490d-86ef-57027d97f282-kube-api-access-xgrmg\") pod \"646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:41 crc kubenswrapper[4573]: I1203 08:55:41.699255 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:42 crc kubenswrapper[4573]: I1203 08:55:42.232984 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp"] Dec 03 08:55:42 crc kubenswrapper[4573]: I1203 08:55:42.795234 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" event={"ID":"c09528b8-a5a1-490d-86ef-57027d97f282","Type":"ContainerStarted","Data":"ead40b2a2ee376e120e057d8054dee937874adf1bf4d0556379a9968b02ca5c0"} Dec 03 08:55:44 crc kubenswrapper[4573]: I1203 08:55:44.819383 4573 generic.go:334] "Generic (PLEG): container finished" podID="c09528b8-a5a1-490d-86ef-57027d97f282" containerID="bc0a0cb1faf9c60ebeaf8e4a3af4929b3f4f35e5cec22212660c542bed41e7ea" exitCode=0 Dec 03 08:55:44 crc kubenswrapper[4573]: I1203 08:55:44.819520 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" event={"ID":"c09528b8-a5a1-490d-86ef-57027d97f282","Type":"ContainerDied","Data":"bc0a0cb1faf9c60ebeaf8e4a3af4929b3f4f35e5cec22212660c542bed41e7ea"} Dec 03 08:55:45 crc kubenswrapper[4573]: I1203 08:55:45.834923 4573 generic.go:334] "Generic (PLEG): container finished" podID="c09528b8-a5a1-490d-86ef-57027d97f282" containerID="8bf473f09cb2294ed6b20bab52e7915b176460550d36e790cb2a463724552295" exitCode=0 Dec 03 08:55:45 crc kubenswrapper[4573]: I1203 08:55:45.835097 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" event={"ID":"c09528b8-a5a1-490d-86ef-57027d97f282","Type":"ContainerDied","Data":"8bf473f09cb2294ed6b20bab52e7915b176460550d36e790cb2a463724552295"} Dec 03 08:55:46 crc kubenswrapper[4573]: I1203 08:55:46.845147 4573 generic.go:334] "Generic (PLEG): container finished" podID="c09528b8-a5a1-490d-86ef-57027d97f282" containerID="5bfeec7937f7902caaa6c277e99d8a3eafdc0708148300c5b94757f8f33ea3eb" exitCode=0 Dec 03 08:55:46 crc kubenswrapper[4573]: I1203 08:55:46.845233 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" event={"ID":"c09528b8-a5a1-490d-86ef-57027d97f282","Type":"ContainerDied","Data":"5bfeec7937f7902caaa6c277e99d8a3eafdc0708148300c5b94757f8f33ea3eb"} Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.124833 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.202316 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgrmg\" (UniqueName: \"kubernetes.io/projected/c09528b8-a5a1-490d-86ef-57027d97f282-kube-api-access-xgrmg\") pod \"c09528b8-a5a1-490d-86ef-57027d97f282\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.202436 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-util\") pod \"c09528b8-a5a1-490d-86ef-57027d97f282\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.202474 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-bundle\") pod \"c09528b8-a5a1-490d-86ef-57027d97f282\" (UID: \"c09528b8-a5a1-490d-86ef-57027d97f282\") " Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.203610 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-bundle" (OuterVolumeSpecName: "bundle") pod "c09528b8-a5a1-490d-86ef-57027d97f282" (UID: "c09528b8-a5a1-490d-86ef-57027d97f282"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.210665 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c09528b8-a5a1-490d-86ef-57027d97f282-kube-api-access-xgrmg" (OuterVolumeSpecName: "kube-api-access-xgrmg") pod "c09528b8-a5a1-490d-86ef-57027d97f282" (UID: "c09528b8-a5a1-490d-86ef-57027d97f282"). InnerVolumeSpecName "kube-api-access-xgrmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.219411 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-util" (OuterVolumeSpecName: "util") pod "c09528b8-a5a1-490d-86ef-57027d97f282" (UID: "c09528b8-a5a1-490d-86ef-57027d97f282"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.304412 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgrmg\" (UniqueName: \"kubernetes.io/projected/c09528b8-a5a1-490d-86ef-57027d97f282-kube-api-access-xgrmg\") on node \"crc\" DevicePath \"\"" Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.304466 4573 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-util\") on node \"crc\" DevicePath \"\"" Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.304481 4573 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c09528b8-a5a1-490d-86ef-57027d97f282-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.865710 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" event={"ID":"c09528b8-a5a1-490d-86ef-57027d97f282","Type":"ContainerDied","Data":"ead40b2a2ee376e120e057d8054dee937874adf1bf4d0556379a9968b02ca5c0"} Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.865777 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ead40b2a2ee376e120e057d8054dee937874adf1bf4d0556379a9968b02ca5c0" Dec 03 08:55:48 crc kubenswrapper[4573]: I1203 08:55:48.865921 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.607168 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr"] Dec 03 08:55:53 crc kubenswrapper[4573]: E1203 08:55:53.608097 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c09528b8-a5a1-490d-86ef-57027d97f282" containerName="pull" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.608117 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c09528b8-a5a1-490d-86ef-57027d97f282" containerName="pull" Dec 03 08:55:53 crc kubenswrapper[4573]: E1203 08:55:53.608137 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c09528b8-a5a1-490d-86ef-57027d97f282" containerName="extract" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.608146 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c09528b8-a5a1-490d-86ef-57027d97f282" containerName="extract" Dec 03 08:55:53 crc kubenswrapper[4573]: E1203 08:55:53.608157 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c09528b8-a5a1-490d-86ef-57027d97f282" containerName="util" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.608164 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c09528b8-a5a1-490d-86ef-57027d97f282" containerName="util" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.608320 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="c09528b8-a5a1-490d-86ef-57027d97f282" containerName="extract" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.608904 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.612655 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-gm9ps" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.709168 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd2jf\" (UniqueName: \"kubernetes.io/projected/42fe335f-c3e5-4792-a174-2f5aeaa7d360-kube-api-access-jd2jf\") pod \"openstack-operator-controller-operator-f6f67fdf8-vgsnr\" (UID: \"42fe335f-c3e5-4792-a174-2f5aeaa7d360\") " pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.717418 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr"] Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.810726 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd2jf\" (UniqueName: \"kubernetes.io/projected/42fe335f-c3e5-4792-a174-2f5aeaa7d360-kube-api-access-jd2jf\") pod \"openstack-operator-controller-operator-f6f67fdf8-vgsnr\" (UID: \"42fe335f-c3e5-4792-a174-2f5aeaa7d360\") " pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.839005 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd2jf\" (UniqueName: \"kubernetes.io/projected/42fe335f-c3e5-4792-a174-2f5aeaa7d360-kube-api-access-jd2jf\") pod \"openstack-operator-controller-operator-f6f67fdf8-vgsnr\" (UID: \"42fe335f-c3e5-4792-a174-2f5aeaa7d360\") " pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" Dec 03 08:55:53 crc kubenswrapper[4573]: I1203 08:55:53.932213 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" Dec 03 08:55:54 crc kubenswrapper[4573]: I1203 08:55:54.268130 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr"] Dec 03 08:55:54 crc kubenswrapper[4573]: I1203 08:55:54.915644 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" event={"ID":"42fe335f-c3e5-4792-a174-2f5aeaa7d360","Type":"ContainerStarted","Data":"26d5a7dcece15ba760d4fe3767a898555c4ecd783609d9b3bc3997d5f8789a21"} Dec 03 08:56:03 crc kubenswrapper[4573]: I1203 08:56:03.092713 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" event={"ID":"42fe335f-c3e5-4792-a174-2f5aeaa7d360","Type":"ContainerStarted","Data":"97f378de7c3bb8cea2df58ebc6f3350dd6197edb3821656c3539b1b8e9e7734b"} Dec 03 08:56:03 crc kubenswrapper[4573]: I1203 08:56:03.093357 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" Dec 03 08:56:03 crc kubenswrapper[4573]: I1203 08:56:03.130477 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" podStartSLOduration=1.6774205119999999 podStartE2EDuration="10.130451444s" podCreationTimestamp="2025-12-03 08:55:53 +0000 UTC" firstStartedPulling="2025-12-03 08:55:54.288626699 +0000 UTC m=+1074.857005958" lastFinishedPulling="2025-12-03 08:56:02.741657631 +0000 UTC m=+1083.310036890" observedRunningTime="2025-12-03 08:56:03.121221343 +0000 UTC m=+1083.689600632" watchObservedRunningTime="2025-12-03 08:56:03.130451444 +0000 UTC m=+1083.698830703" Dec 03 08:56:13 crc kubenswrapper[4573]: I1203 08:56:13.935768 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-f6f67fdf8-vgsnr" Dec 03 08:56:33 crc kubenswrapper[4573]: I1203 08:56:33.957217 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq"] Dec 03 08:56:33 crc kubenswrapper[4573]: I1203 08:56:33.959554 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" Dec 03 08:56:33 crc kubenswrapper[4573]: I1203 08:56:33.969775 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-s89j5" Dec 03 08:56:33 crc kubenswrapper[4573]: I1203 08:56:33.986725 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq"] Dec 03 08:56:33 crc kubenswrapper[4573]: I1203 08:56:33.991223 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4"] Dec 03 08:56:33 crc kubenswrapper[4573]: I1203 08:56:33.994751 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.005948 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-t22gm" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.015040 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.062348 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msvwb\" (UniqueName: \"kubernetes.io/projected/e27aea95-097f-4aa8-ae2b-7d212b8640b4-kube-api-access-msvwb\") pod \"barbican-operator-controller-manager-7d9dfd778-nx7mq\" (UID: \"e27aea95-097f-4aa8-ae2b-7d212b8640b4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.062497 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kknkc\" (UniqueName: \"kubernetes.io/projected/e0ae572b-e68c-48a0-8649-7eea884e61b0-kube-api-access-kknkc\") pod \"cinder-operator-controller-manager-859b6ccc6-hb6l4\" (UID: \"e0ae572b-e68c-48a0-8649-7eea884e61b0\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.079242 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.080917 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.088791 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-qdgv8" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.104852 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.106198 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.111890 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.114472 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-9npp8" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.123105 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.124297 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.139130 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.139754 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-d4nhn" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.163238 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxwr2\" (UniqueName: \"kubernetes.io/projected/ee80479f-97a6-43ef-a026-d5bf1931d962-kube-api-access-rxwr2\") pod \"designate-operator-controller-manager-78b4bc895b-vl9xb\" (UID: \"ee80479f-97a6-43ef-a026-d5bf1931d962\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.163328 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msvwb\" (UniqueName: \"kubernetes.io/projected/e27aea95-097f-4aa8-ae2b-7d212b8640b4-kube-api-access-msvwb\") pod \"barbican-operator-controller-manager-7d9dfd778-nx7mq\" (UID: \"e27aea95-097f-4aa8-ae2b-7d212b8640b4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.163373 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kknkc\" (UniqueName: \"kubernetes.io/projected/e0ae572b-e68c-48a0-8649-7eea884e61b0-kube-api-access-kknkc\") pod \"cinder-operator-controller-manager-859b6ccc6-hb6l4\" (UID: \"e0ae572b-e68c-48a0-8649-7eea884e61b0\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.163395 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wprn\" (UniqueName: \"kubernetes.io/projected/622de9fb-c4ba-4727-8c22-5fcb36f39751-kube-api-access-7wprn\") pod \"horizon-operator-controller-manager-68c6d99b8f-2qnlp\" (UID: \"622de9fb-c4ba-4727-8c22-5fcb36f39751\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.163416 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj544\" (UniqueName: \"kubernetes.io/projected/a855d891-6a41-477d-98d7-fcbba59aee28-kube-api-access-xj544\") pod \"glance-operator-controller-manager-77987cd8cd-s8679\" (UID: \"a855d891-6a41-477d-98d7-fcbba59aee28\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.168430 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.170742 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.178854 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-27686" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.189476 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.194127 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gljjt"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.195432 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.203852 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.203960 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-c5nj8" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.205528 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msvwb\" (UniqueName: \"kubernetes.io/projected/e27aea95-097f-4aa8-ae2b-7d212b8640b4-kube-api-access-msvwb\") pod \"barbican-operator-controller-manager-7d9dfd778-nx7mq\" (UID: \"e27aea95-097f-4aa8-ae2b-7d212b8640b4\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.209372 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kknkc\" (UniqueName: \"kubernetes.io/projected/e0ae572b-e68c-48a0-8649-7eea884e61b0-kube-api-access-kknkc\") pod \"cinder-operator-controller-manager-859b6ccc6-hb6l4\" (UID: \"e0ae572b-e68c-48a0-8649-7eea884e61b0\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.263149 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.264790 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wprn\" (UniqueName: \"kubernetes.io/projected/622de9fb-c4ba-4727-8c22-5fcb36f39751-kube-api-access-7wprn\") pod \"horizon-operator-controller-manager-68c6d99b8f-2qnlp\" (UID: \"622de9fb-c4ba-4727-8c22-5fcb36f39751\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.264976 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj544\" (UniqueName: \"kubernetes.io/projected/a855d891-6a41-477d-98d7-fcbba59aee28-kube-api-access-xj544\") pod \"glance-operator-controller-manager-77987cd8cd-s8679\" (UID: \"a855d891-6a41-477d-98d7-fcbba59aee28\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.265162 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg7hv\" (UniqueName: \"kubernetes.io/projected/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-kube-api-access-dg7hv\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.265315 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxwr2\" (UniqueName: \"kubernetes.io/projected/ee80479f-97a6-43ef-a026-d5bf1931d962-kube-api-access-rxwr2\") pod \"designate-operator-controller-manager-78b4bc895b-vl9xb\" (UID: \"ee80479f-97a6-43ef-a026-d5bf1931d962\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.265422 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.265538 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x97tw\" (UniqueName: \"kubernetes.io/projected/ff1aaf12-2fcd-41d0-a575-a0e5447fc11f-kube-api-access-x97tw\") pod \"heat-operator-controller-manager-5f64f6f8bb-gnlgf\" (UID: \"ff1aaf12-2fcd-41d0-a575-a0e5447fc11f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.271388 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.272780 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.291257 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tvtfd" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.296437 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gljjt"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.302593 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.314560 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.341408 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.363200 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.364819 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wprn\" (UniqueName: \"kubernetes.io/projected/622de9fb-c4ba-4727-8c22-5fcb36f39751-kube-api-access-7wprn\") pod \"horizon-operator-controller-manager-68c6d99b8f-2qnlp\" (UID: \"622de9fb-c4ba-4727-8c22-5fcb36f39751\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.368101 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4lrl\" (UniqueName: \"kubernetes.io/projected/5aba2f13-92bd-4ea3-bfcb-7646909db04b-kube-api-access-p4lrl\") pod \"keystone-operator-controller-manager-7765d96ddf-hdhrp\" (UID: \"5aba2f13-92bd-4ea3-bfcb-7646909db04b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.368152 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7cfd\" (UniqueName: \"kubernetes.io/projected/9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b-kube-api-access-v7cfd\") pod \"ironic-operator-controller-manager-6c548fd776-wcbtn\" (UID: \"9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.368218 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg7hv\" (UniqueName: \"kubernetes.io/projected/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-kube-api-access-dg7hv\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.368270 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.368296 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x97tw\" (UniqueName: \"kubernetes.io/projected/ff1aaf12-2fcd-41d0-a575-a0e5447fc11f-kube-api-access-x97tw\") pod \"heat-operator-controller-manager-5f64f6f8bb-gnlgf\" (UID: \"ff1aaf12-2fcd-41d0-a575-a0e5447fc11f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" Dec 03 08:56:34 crc kubenswrapper[4573]: E1203 08:56:34.368971 4573 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:34 crc kubenswrapper[4573]: E1203 08:56:34.369036 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert podName:22bca7e1-22a7-4ee7-852b-25fddeb8fce3 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:34.869011376 +0000 UTC m=+1115.437390635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert") pod "infra-operator-controller-manager-57548d458d-gljjt" (UID: "22bca7e1-22a7-4ee7-852b-25fddeb8fce3") : secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.378373 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.385700 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxwr2\" (UniqueName: \"kubernetes.io/projected/ee80479f-97a6-43ef-a026-d5bf1931d962-kube-api-access-rxwr2\") pod \"designate-operator-controller-manager-78b4bc895b-vl9xb\" (UID: \"ee80479f-97a6-43ef-a026-d5bf1931d962\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.393721 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj544\" (UniqueName: \"kubernetes.io/projected/a855d891-6a41-477d-98d7-fcbba59aee28-kube-api-access-xj544\") pod \"glance-operator-controller-manager-77987cd8cd-s8679\" (UID: \"a855d891-6a41-477d-98d7-fcbba59aee28\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.394117 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-cplhn" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.406580 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.432213 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.441982 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.465956 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.467124 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x97tw\" (UniqueName: \"kubernetes.io/projected/ff1aaf12-2fcd-41d0-a575-a0e5447fc11f-kube-api-access-x97tw\") pod \"heat-operator-controller-manager-5f64f6f8bb-gnlgf\" (UID: \"ff1aaf12-2fcd-41d0-a575-a0e5447fc11f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.470573 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4lrl\" (UniqueName: \"kubernetes.io/projected/5aba2f13-92bd-4ea3-bfcb-7646909db04b-kube-api-access-p4lrl\") pod \"keystone-operator-controller-manager-7765d96ddf-hdhrp\" (UID: \"5aba2f13-92bd-4ea3-bfcb-7646909db04b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.470609 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7cfd\" (UniqueName: \"kubernetes.io/projected/9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b-kube-api-access-v7cfd\") pod \"ironic-operator-controller-manager-6c548fd776-wcbtn\" (UID: \"9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.487972 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.489500 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.501658 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.505249 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg7hv\" (UniqueName: \"kubernetes.io/projected/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-kube-api-access-dg7hv\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.505506 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-2cc78" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.506263 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.556677 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7cfd\" (UniqueName: \"kubernetes.io/projected/9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b-kube-api-access-v7cfd\") pod \"ironic-operator-controller-manager-6c548fd776-wcbtn\" (UID: \"9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.568615 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4lrl\" (UniqueName: \"kubernetes.io/projected/5aba2f13-92bd-4ea3-bfcb-7646909db04b-kube-api-access-p4lrl\") pod \"keystone-operator-controller-manager-7765d96ddf-hdhrp\" (UID: \"5aba2f13-92bd-4ea3-bfcb-7646909db04b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.581058 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr8r5\" (UniqueName: \"kubernetes.io/projected/4613c040-3ca1-458c-8ff8-2b8858a7ad35-kube-api-access-pr8r5\") pod \"manila-operator-controller-manager-7c79b5df47-qphvx\" (UID: \"4613c040-3ca1-458c-8ff8-2b8858a7ad35\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.616653 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.617923 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.635596 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-mllgl" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.679326 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-bnztb"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.680489 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.684706 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d44qp\" (UniqueName: \"kubernetes.io/projected/a2888fa7-1f19-48c8-853e-a3952cc93ec6-kube-api-access-d44qp\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hl72w\" (UID: \"a2888fa7-1f19-48c8-853e-a3952cc93ec6\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.684809 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr8r5\" (UniqueName: \"kubernetes.io/projected/4613c040-3ca1-458c-8ff8-2b8858a7ad35-kube-api-access-pr8r5\") pod \"manila-operator-controller-manager-7c79b5df47-qphvx\" (UID: \"4613c040-3ca1-458c-8ff8-2b8858a7ad35\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.732145 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-kpmj9" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.757507 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.772288 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.781453 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.783005 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.791191 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-n42v8" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.792720 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d44qp\" (UniqueName: \"kubernetes.io/projected/a2888fa7-1f19-48c8-853e-a3952cc93ec6-kube-api-access-d44qp\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hl72w\" (UID: \"a2888fa7-1f19-48c8-853e-a3952cc93ec6\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.808337 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.816877 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.817841 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.828944 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-2l7q7" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.877622 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-bnztb"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.893400 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr8r5\" (UniqueName: \"kubernetes.io/projected/4613c040-3ca1-458c-8ff8-2b8858a7ad35-kube-api-access-pr8r5\") pod \"manila-operator-controller-manager-7c79b5df47-qphvx\" (UID: \"4613c040-3ca1-458c-8ff8-2b8858a7ad35\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.894208 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.894253 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d6pk\" (UniqueName: \"kubernetes.io/projected/fd426f33-5f03-4fec-bde4-e72e9c762762-kube-api-access-5d6pk\") pod \"nova-operator-controller-manager-697bc559fc-sdrbz\" (UID: \"fd426f33-5f03-4fec-bde4-e72e9c762762\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.894279 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgdl8\" (UniqueName: \"kubernetes.io/projected/98fcc67c-249b-4700-8434-c8dd4e1646ae-kube-api-access-cgdl8\") pod \"octavia-operator-controller-manager-998648c74-bnztb\" (UID: \"98fcc67c-249b-4700-8434-c8dd4e1646ae\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" Dec 03 08:56:34 crc kubenswrapper[4573]: E1203 08:56:34.894448 4573 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:34 crc kubenswrapper[4573]: E1203 08:56:34.894501 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert podName:22bca7e1-22a7-4ee7-852b-25fddeb8fce3 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:35.894484119 +0000 UTC m=+1116.462863378 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert") pod "infra-operator-controller-manager-57548d458d-gljjt" (UID: "22bca7e1-22a7-4ee7-852b-25fddeb8fce3") : secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.897903 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.907304 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb"] Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.910902 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d44qp\" (UniqueName: \"kubernetes.io/projected/a2888fa7-1f19-48c8-853e-a3952cc93ec6-kube-api-access-d44qp\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hl72w\" (UID: \"a2888fa7-1f19-48c8-853e-a3952cc93ec6\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.955428 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.996782 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgmfj\" (UniqueName: \"kubernetes.io/projected/9d6a307e-5cf2-4a23-921c-c5b562494cb0-kube-api-access-sgmfj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-r6qvb\" (UID: \"9d6a307e-5cf2-4a23-921c-c5b562494cb0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.996887 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d6pk\" (UniqueName: \"kubernetes.io/projected/fd426f33-5f03-4fec-bde4-e72e9c762762-kube-api-access-5d6pk\") pod \"nova-operator-controller-manager-697bc559fc-sdrbz\" (UID: \"fd426f33-5f03-4fec-bde4-e72e9c762762\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" Dec 03 08:56:34 crc kubenswrapper[4573]: I1203 08:56:34.996919 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgdl8\" (UniqueName: \"kubernetes.io/projected/98fcc67c-249b-4700-8434-c8dd4e1646ae-kube-api-access-cgdl8\") pod \"octavia-operator-controller-manager-998648c74-bnztb\" (UID: \"98fcc67c-249b-4700-8434-c8dd4e1646ae\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.108114 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.111099 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgmfj\" (UniqueName: \"kubernetes.io/projected/9d6a307e-5cf2-4a23-921c-c5b562494cb0-kube-api-access-sgmfj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-r6qvb\" (UID: \"9d6a307e-5cf2-4a23-921c-c5b562494cb0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.112394 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.134444 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-x5txt" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.134632 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.135208 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.152084 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d6pk\" (UniqueName: \"kubernetes.io/projected/fd426f33-5f03-4fec-bde4-e72e9c762762-kube-api-access-5d6pk\") pod \"nova-operator-controller-manager-697bc559fc-sdrbz\" (UID: \"fd426f33-5f03-4fec-bde4-e72e9c762762\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.175725 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.189518 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgdl8\" (UniqueName: \"kubernetes.io/projected/98fcc67c-249b-4700-8434-c8dd4e1646ae-kube-api-access-cgdl8\") pod \"octavia-operator-controller-manager-998648c74-bnztb\" (UID: \"98fcc67c-249b-4700-8434-c8dd4e1646ae\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.212285 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t52t\" (UniqueName: \"kubernetes.io/projected/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-kube-api-access-5t52t\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.212330 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.228736 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.229129 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.247007 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.290256 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-mpb8m" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.334253 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t52t\" (UniqueName: \"kubernetes.io/projected/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-kube-api-access-5t52t\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.334306 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.334420 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7wxm\" (UniqueName: \"kubernetes.io/projected/fd5e6711-117e-42b2-90e6-cfddf410315d-kube-api-access-z7wxm\") pod \"ovn-operator-controller-manager-b6456fdb6-d77h9\" (UID: \"fd5e6711-117e-42b2-90e6-cfddf410315d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" Dec 03 08:56:35 crc kubenswrapper[4573]: E1203 08:56:35.334747 4573 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:35 crc kubenswrapper[4573]: E1203 08:56:35.334793 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert podName:a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:35.8347779 +0000 UTC m=+1116.403157159 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" (UID: "a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.354676 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.356183 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.359299 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgmfj\" (UniqueName: \"kubernetes.io/projected/9d6a307e-5cf2-4a23-921c-c5b562494cb0-kube-api-access-sgmfj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-r6qvb\" (UID: \"9d6a307e-5cf2-4a23-921c-c5b562494cb0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.361686 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.368502 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-8vvsl" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.368945 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.370333 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.384395 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-4tgw7" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.406680 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.415295 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.467849 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.470771 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5txz\" (UniqueName: \"kubernetes.io/projected/e94079e6-8701-48bf-ab58-1867fdf6e46e-kube-api-access-p5txz\") pod \"placement-operator-controller-manager-78f8948974-c5vc7\" (UID: \"e94079e6-8701-48bf-ab58-1867fdf6e46e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.470834 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7wxm\" (UniqueName: \"kubernetes.io/projected/fd5e6711-117e-42b2-90e6-cfddf410315d-kube-api-access-z7wxm\") pod \"ovn-operator-controller-manager-b6456fdb6-d77h9\" (UID: \"fd5e6711-117e-42b2-90e6-cfddf410315d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.470888 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp9wk\" (UniqueName: \"kubernetes.io/projected/57f767e9-7c3a-4707-9b76-ed715c297238-kube-api-access-jp9wk\") pod \"swift-operator-controller-manager-5f8c65bbfc-nbfxj\" (UID: \"57f767e9-7c3a-4707-9b76-ed715c297238\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.472295 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t52t\" (UniqueName: \"kubernetes.io/projected/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-kube-api-access-5t52t\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.567282 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7wxm\" (UniqueName: \"kubernetes.io/projected/fd5e6711-117e-42b2-90e6-cfddf410315d-kube-api-access-z7wxm\") pod \"ovn-operator-controller-manager-b6456fdb6-d77h9\" (UID: \"fd5e6711-117e-42b2-90e6-cfddf410315d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.616793 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5txz\" (UniqueName: \"kubernetes.io/projected/e94079e6-8701-48bf-ab58-1867fdf6e46e-kube-api-access-p5txz\") pod \"placement-operator-controller-manager-78f8948974-c5vc7\" (UID: \"e94079e6-8701-48bf-ab58-1867fdf6e46e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.616926 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp9wk\" (UniqueName: \"kubernetes.io/projected/57f767e9-7c3a-4707-9b76-ed715c297238-kube-api-access-jp9wk\") pod \"swift-operator-controller-manager-5f8c65bbfc-nbfxj\" (UID: \"57f767e9-7c3a-4707-9b76-ed715c297238\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.677322 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5txz\" (UniqueName: \"kubernetes.io/projected/e94079e6-8701-48bf-ab58-1867fdf6e46e-kube-api-access-p5txz\") pod \"placement-operator-controller-manager-78f8948974-c5vc7\" (UID: \"e94079e6-8701-48bf-ab58-1867fdf6e46e\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.686518 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.703179 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.716442 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.722414 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-lrl22" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.798579 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp9wk\" (UniqueName: \"kubernetes.io/projected/57f767e9-7c3a-4707-9b76-ed715c297238-kube-api-access-jp9wk\") pod \"swift-operator-controller-manager-5f8c65bbfc-nbfxj\" (UID: \"57f767e9-7c3a-4707-9b76-ed715c297238\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.821980 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c2vt\" (UniqueName: \"kubernetes.io/projected/8f65995e-e70a-48e4-9504-8aa85d43ea9e-kube-api-access-2c2vt\") pod \"telemetry-operator-controller-manager-76cc84c6bb-jd6bv\" (UID: \"8f65995e-e70a-48e4-9504-8aa85d43ea9e\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.822109 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.822644 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.868798 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj"] Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.870493 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.877530 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-rz94r" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.923704 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.923768 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c2vt\" (UniqueName: \"kubernetes.io/projected/8f65995e-e70a-48e4-9504-8aa85d43ea9e-kube-api-access-2c2vt\") pod \"telemetry-operator-controller-manager-76cc84c6bb-jd6bv\" (UID: \"8f65995e-e70a-48e4-9504-8aa85d43ea9e\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.923877 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:35 crc kubenswrapper[4573]: E1203 08:56:35.924032 4573 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:35 crc kubenswrapper[4573]: E1203 08:56:35.924133 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert podName:22bca7e1-22a7-4ee7-852b-25fddeb8fce3 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:37.924114232 +0000 UTC m=+1118.492493491 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert") pod "infra-operator-controller-manager-57548d458d-gljjt" (UID: "22bca7e1-22a7-4ee7-852b-25fddeb8fce3") : secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:35 crc kubenswrapper[4573]: E1203 08:56:35.924602 4573 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:35 crc kubenswrapper[4573]: E1203 08:56:35.935709 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert podName:a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:36.935661496 +0000 UTC m=+1117.504040755 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" (UID: "a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:35 crc kubenswrapper[4573]: I1203 08:56:35.963816 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.000632 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.022071 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c2vt\" (UniqueName: \"kubernetes.io/projected/8f65995e-e70a-48e4-9504-8aa85d43ea9e-kube-api-access-2c2vt\") pod \"telemetry-operator-controller-manager-76cc84c6bb-jd6bv\" (UID: \"8f65995e-e70a-48e4-9504-8aa85d43ea9e\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.028517 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdd56\" (UniqueName: \"kubernetes.io/projected/6535c22e-f476-4067-b1ab-7e0ed5c23360-kube-api-access-sdd56\") pod \"test-operator-controller-manager-5854674fcc-wdvlj\" (UID: \"6535c22e-f476-4067-b1ab-7e0ed5c23360\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.072087 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.110402 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.111876 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.111964 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.135086 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdd56\" (UniqueName: \"kubernetes.io/projected/6535c22e-f476-4067-b1ab-7e0ed5c23360-kube-api-access-sdd56\") pod \"test-operator-controller-manager-5854674fcc-wdvlj\" (UID: \"6535c22e-f476-4067-b1ab-7e0ed5c23360\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.144690 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-c49m5" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.178468 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.191788 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdd56\" (UniqueName: \"kubernetes.io/projected/6535c22e-f476-4067-b1ab-7e0ed5c23360-kube-api-access-sdd56\") pod \"test-operator-controller-manager-5854674fcc-wdvlj\" (UID: \"6535c22e-f476-4067-b1ab-7e0ed5c23360\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.237979 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh2km\" (UniqueName: \"kubernetes.io/projected/e6f9febf-6db7-4ea8-8b96-7a939102be2a-kube-api-access-bh2km\") pod \"watcher-operator-controller-manager-769dc69bc-l5vf4\" (UID: \"e6f9febf-6db7-4ea8-8b96-7a939102be2a\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.280826 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.314862 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.316145 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.324599 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.334725 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-8cdnp" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.334936 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.335449 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.345680 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh2km\" (UniqueName: \"kubernetes.io/projected/e6f9febf-6db7-4ea8-8b96-7a939102be2a-kube-api-access-bh2km\") pod \"watcher-operator-controller-manager-769dc69bc-l5vf4\" (UID: \"e6f9febf-6db7-4ea8-8b96-7a939102be2a\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.367231 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.401287 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh2km\" (UniqueName: \"kubernetes.io/projected/e6f9febf-6db7-4ea8-8b96-7a939102be2a-kube-api-access-bh2km\") pod \"watcher-operator-controller-manager-769dc69bc-l5vf4\" (UID: \"e6f9febf-6db7-4ea8-8b96-7a939102be2a\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.428401 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.428539 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.442411 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq"] Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.448367 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.448453 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.448520 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwhrw\" (UniqueName: \"kubernetes.io/projected/5acea32c-fc74-40d8-b363-c64470ee880e-kube-api-access-fwhrw\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:36 crc kubenswrapper[4573]: W1203 08:56:36.452562 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode27aea95_097f_4aa8_ae2b_7d212b8640b4.slice/crio-715cb3de6978067fa79d082c3a68c7b0366a35d9f5d5e2d5edb986f0b2d951b0 WatchSource:0}: Error finding container 715cb3de6978067fa79d082c3a68c7b0366a35d9f5d5e2d5edb986f0b2d951b0: Status 404 returned error can't find the container with id 715cb3de6978067fa79d082c3a68c7b0366a35d9f5d5e2d5edb986f0b2d951b0 Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.454059 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-mmxq2" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.553473 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.586278 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwhrw\" (UniqueName: \"kubernetes.io/projected/5acea32c-fc74-40d8-b363-c64470ee880e-kube-api-access-fwhrw\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.586536 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.586694 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.586811 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gx5k\" (UniqueName: \"kubernetes.io/projected/b4436129-0e91-443b-8a97-61fa7f8cc2e4-kube-api-access-9gx5k\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jbwlq\" (UID: \"b4436129-0e91-443b-8a97-61fa7f8cc2e4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" Dec 03 08:56:36 crc kubenswrapper[4573]: E1203 08:56:36.587989 4573 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 08:56:36 crc kubenswrapper[4573]: E1203 08:56:36.588065 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:37.088025324 +0000 UTC m=+1117.656404583 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "metrics-server-cert" not found Dec 03 08:56:36 crc kubenswrapper[4573]: E1203 08:56:36.588268 4573 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 08:56:36 crc kubenswrapper[4573]: E1203 08:56:36.588300 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:37.088290491 +0000 UTC m=+1117.656669750 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "webhook-server-cert" not found Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.632133 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwhrw\" (UniqueName: \"kubernetes.io/projected/5acea32c-fc74-40d8-b363-c64470ee880e-kube-api-access-fwhrw\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.690249 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gx5k\" (UniqueName: \"kubernetes.io/projected/b4436129-0e91-443b-8a97-61fa7f8cc2e4-kube-api-access-9gx5k\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jbwlq\" (UID: \"b4436129-0e91-443b-8a97-61fa7f8cc2e4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.722802 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gx5k\" (UniqueName: \"kubernetes.io/projected/b4436129-0e91-443b-8a97-61fa7f8cc2e4-kube-api-access-9gx5k\") pod \"rabbitmq-cluster-operator-manager-668c99d594-jbwlq\" (UID: \"b4436129-0e91-443b-8a97-61fa7f8cc2e4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" Dec 03 08:56:36 crc kubenswrapper[4573]: I1203 08:56:36.838676 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.000883 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:37 crc kubenswrapper[4573]: E1203 08:56:37.001142 4573 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:37 crc kubenswrapper[4573]: E1203 08:56:37.001191 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert podName:a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:39.001176788 +0000 UTC m=+1119.569556047 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" (UID: "a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.105339 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.105506 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:37 crc kubenswrapper[4573]: E1203 08:56:37.105657 4573 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 08:56:37 crc kubenswrapper[4573]: E1203 08:56:37.105693 4573 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 08:56:37 crc kubenswrapper[4573]: E1203 08:56:37.105772 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:38.105741935 +0000 UTC m=+1118.674121194 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "webhook-server-cert" not found Dec 03 08:56:37 crc kubenswrapper[4573]: E1203 08:56:37.105796 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:38.105788957 +0000 UTC m=+1118.674168216 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "metrics-server-cert" not found Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.227178 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp"] Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.243292 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb"] Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.328709 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4"] Dec 03 08:56:37 crc kubenswrapper[4573]: W1203 08:56:37.401471 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0ae572b_e68c_48a0_8649_7eea884e61b0.slice/crio-a566e09b3f4d929ed6cdf9b2935512452d8c66712426937249b8677f0a941e7e WatchSource:0}: Error finding container a566e09b3f4d929ed6cdf9b2935512452d8c66712426937249b8677f0a941e7e: Status 404 returned error can't find the container with id a566e09b3f4d929ed6cdf9b2935512452d8c66712426937249b8677f0a941e7e Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.461835 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf"] Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.573953 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-bnztb"] Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.580436 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" event={"ID":"e27aea95-097f-4aa8-ae2b-7d212b8640b4","Type":"ContainerStarted","Data":"715cb3de6978067fa79d082c3a68c7b0366a35d9f5d5e2d5edb986f0b2d951b0"} Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.625931 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn"] Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.679543 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679"] Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.679599 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" event={"ID":"ff1aaf12-2fcd-41d0-a575-a0e5447fc11f","Type":"ContainerStarted","Data":"1150d4933c9355ae6f15816d869ac3dde84d73748e8731cbfe6ffab95af001f4"} Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.703274 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w"] Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.733449 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" event={"ID":"ee80479f-97a6-43ef-a026-d5bf1931d962","Type":"ContainerStarted","Data":"a982b982fa91b7e844419aff6725d137ac41949846edc1ce2dc696ca2a704f7d"} Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.762994 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" event={"ID":"622de9fb-c4ba-4727-8c22-5fcb36f39751","Type":"ContainerStarted","Data":"e6edd98f1e4cd90c3ccd6d559ecd01ce02b7bb38076ea1afa19e918816a4f323"} Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.798782 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" event={"ID":"e0ae572b-e68c-48a0-8649-7eea884e61b0","Type":"ContainerStarted","Data":"a566e09b3f4d929ed6cdf9b2935512452d8c66712426937249b8677f0a941e7e"} Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.863150 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp"] Dec 03 08:56:37 crc kubenswrapper[4573]: I1203 08:56:37.945236 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:37 crc kubenswrapper[4573]: E1203 08:56:37.945401 4573 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:37 crc kubenswrapper[4573]: E1203 08:56:37.945594 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert podName:22bca7e1-22a7-4ee7-852b-25fddeb8fce3 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:41.94557779 +0000 UTC m=+1122.513957059 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert") pod "infra-operator-controller-manager-57548d458d-gljjt" (UID: "22bca7e1-22a7-4ee7-852b-25fddeb8fce3") : secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.094918 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7"] Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.094975 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx"] Dec 03 08:56:38 crc kubenswrapper[4573]: W1203 08:56:38.113199 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4613c040_3ca1_458c_8ff8_2b8858a7ad35.slice/crio-ae5774618b5d8e79610c0aca13b5920bf15271597f657d86744db7692f014f15 WatchSource:0}: Error finding container ae5774618b5d8e79610c0aca13b5920bf15271597f657d86744db7692f014f15: Status 404 returned error can't find the container with id ae5774618b5d8e79610c0aca13b5920bf15271597f657d86744db7692f014f15 Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.149215 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.149351 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.149516 4573 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.149585 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:40.149564636 +0000 UTC m=+1120.717943895 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "metrics-server-cert" not found Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.149638 4573 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.149660 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:40.149653858 +0000 UTC m=+1120.718033117 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "webhook-server-cert" not found Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.216877 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz"] Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.255750 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv"] Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.334757 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb"] Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.397789 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9"] Dec 03 08:56:38 crc kubenswrapper[4573]: W1203 08:56:38.401535 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d6a307e_5cf2_4a23_921c_c5b562494cb0.slice/crio-d250e1054bb1aeeb38c7406e667545bf70459a84f1a1c5a34565c440feb47cf5 WatchSource:0}: Error finding container d250e1054bb1aeeb38c7406e667545bf70459a84f1a1c5a34565c440feb47cf5: Status 404 returned error can't find the container with id d250e1054bb1aeeb38c7406e667545bf70459a84f1a1c5a34565c440feb47cf5 Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.465115 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj"] Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.480114 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4"] Dec 03 08:56:38 crc kubenswrapper[4573]: W1203 08:56:38.531318 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6535c22e_f476_4067_b1ab_7e0ed5c23360.slice/crio-36b2c8505ce1a16263e009635270ca7b9f4ce20aad48cab44e006786c794c513 WatchSource:0}: Error finding container 36b2c8505ce1a16263e009635270ca7b9f4ce20aad48cab44e006786c794c513: Status 404 returned error can't find the container with id 36b2c8505ce1a16263e009635270ca7b9f4ce20aad48cab44e006786c794c513 Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.543119 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq"] Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.548056 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj"] Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.588781 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bh2km,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-l5vf4_openstack-operators(e6f9febf-6db7-4ea8-8b96-7a939102be2a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.597352 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bh2km,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-l5vf4_openstack-operators(e6f9febf-6db7-4ea8-8b96-7a939102be2a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.599506 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" podUID="e6f9febf-6db7-4ea8-8b96-7a939102be2a" Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.599648 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9gx5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-jbwlq_openstack-operators(b4436129-0e91-443b-8a97-61fa7f8cc2e4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.601138 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" podUID="b4436129-0e91-443b-8a97-61fa7f8cc2e4" Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.666679 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jp9wk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-nbfxj_openstack-operators(57f767e9-7c3a-4707-9b76-ed715c297238): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.674434 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jp9wk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-nbfxj_openstack-operators(57f767e9-7c3a-4707-9b76-ed715c297238): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.675547 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" podUID="57f767e9-7c3a-4707-9b76-ed715c297238" Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.871580 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" event={"ID":"5aba2f13-92bd-4ea3-bfcb-7646909db04b","Type":"ContainerStarted","Data":"3829bbedc525c2ea329d22d1b413bd7e7fa33ee9534d164674b5b5d8f4ad181d"} Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.876210 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" event={"ID":"8f65995e-e70a-48e4-9504-8aa85d43ea9e","Type":"ContainerStarted","Data":"8fdaa44fc4b421499a02bad1863e195a3349927fa7412e0bbd566c38ef3247b4"} Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.893764 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" event={"ID":"fd5e6711-117e-42b2-90e6-cfddf410315d","Type":"ContainerStarted","Data":"f18ed6183b13570de6b0e1e5bd45c8d5c6e19256563ebd82747cb28fb8cd0bcf"} Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.931795 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" event={"ID":"98fcc67c-249b-4700-8434-c8dd4e1646ae","Type":"ContainerStarted","Data":"5dfa1939cd528df670ae4b25f69a72a9c9ae85e24850eb7ba7750b48338936e4"} Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.945041 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" event={"ID":"57f767e9-7c3a-4707-9b76-ed715c297238","Type":"ContainerStarted","Data":"2223ab90e0444de87630d4a4b96d6005f9c73c321c3b0596dbe242d1280001dd"} Dec 03 08:56:38 crc kubenswrapper[4573]: E1203 08:56:38.964551 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" podUID="57f767e9-7c3a-4707-9b76-ed715c297238" Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.975739 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" event={"ID":"e94079e6-8701-48bf-ab58-1867fdf6e46e","Type":"ContainerStarted","Data":"8c7f4fe377c77e4342afd21d2df731e2901ea80779fbc1a63d5235bf8f5eed92"} Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.989354 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" event={"ID":"9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b","Type":"ContainerStarted","Data":"12a62558a14f7b462daff75c0502f9ab67f3e76496462c46198fbebba7f0736a"} Dec 03 08:56:38 crc kubenswrapper[4573]: I1203 08:56:38.993895 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" event={"ID":"e6f9febf-6db7-4ea8-8b96-7a939102be2a","Type":"ContainerStarted","Data":"edad39634df6b1f4a02d6bd520de69302351be5d887e4914b221e09e877d9814"} Dec 03 08:56:39 crc kubenswrapper[4573]: I1203 08:56:39.000530 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" event={"ID":"6535c22e-f476-4067-b1ab-7e0ed5c23360","Type":"ContainerStarted","Data":"36b2c8505ce1a16263e009635270ca7b9f4ce20aad48cab44e006786c794c513"} Dec 03 08:56:39 crc kubenswrapper[4573]: E1203 08:56:39.004285 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" podUID="e6f9febf-6db7-4ea8-8b96-7a939102be2a" Dec 03 08:56:39 crc kubenswrapper[4573]: I1203 08:56:39.004427 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" event={"ID":"4613c040-3ca1-458c-8ff8-2b8858a7ad35","Type":"ContainerStarted","Data":"ae5774618b5d8e79610c0aca13b5920bf15271597f657d86744db7692f014f15"} Dec 03 08:56:39 crc kubenswrapper[4573]: I1203 08:56:39.007620 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:39 crc kubenswrapper[4573]: E1203 08:56:39.007835 4573 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:39 crc kubenswrapper[4573]: E1203 08:56:39.007887 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert podName:a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:43.007869953 +0000 UTC m=+1123.576249202 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" (UID: "a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:39 crc kubenswrapper[4573]: I1203 08:56:39.011857 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" event={"ID":"fd426f33-5f03-4fec-bde4-e72e9c762762","Type":"ContainerStarted","Data":"46db6d0641b8787c6b6004555952e8e0a7fbd7146bee2aad06016b00c744301f"} Dec 03 08:56:39 crc kubenswrapper[4573]: I1203 08:56:39.025950 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" event={"ID":"a855d891-6a41-477d-98d7-fcbba59aee28","Type":"ContainerStarted","Data":"9a8b9b38083f8f15c4e47fcc17df3a95db3e694f2307af38530e351b7db2c07f"} Dec 03 08:56:39 crc kubenswrapper[4573]: I1203 08:56:39.062776 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" event={"ID":"b4436129-0e91-443b-8a97-61fa7f8cc2e4","Type":"ContainerStarted","Data":"4871e0ec9195d138afd080e3902196a78a1be139aece67d633f5a06a9498a0ea"} Dec 03 08:56:39 crc kubenswrapper[4573]: E1203 08:56:39.066977 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" podUID="b4436129-0e91-443b-8a97-61fa7f8cc2e4" Dec 03 08:56:39 crc kubenswrapper[4573]: I1203 08:56:39.068810 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" event={"ID":"9d6a307e-5cf2-4a23-921c-c5b562494cb0","Type":"ContainerStarted","Data":"d250e1054bb1aeeb38c7406e667545bf70459a84f1a1c5a34565c440feb47cf5"} Dec 03 08:56:39 crc kubenswrapper[4573]: I1203 08:56:39.072960 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" event={"ID":"a2888fa7-1f19-48c8-853e-a3952cc93ec6","Type":"ContainerStarted","Data":"16e7d365042a02c94f4b2acac6b5c9b0584f736bb63b0cc168926e5e6576339a"} Dec 03 08:56:40 crc kubenswrapper[4573]: E1203 08:56:40.191596 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" podUID="b4436129-0e91-443b-8a97-61fa7f8cc2e4" Dec 03 08:56:40 crc kubenswrapper[4573]: E1203 08:56:40.206986 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" podUID="e6f9febf-6db7-4ea8-8b96-7a939102be2a" Dec 03 08:56:40 crc kubenswrapper[4573]: E1203 08:56:40.207522 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" podUID="57f767e9-7c3a-4707-9b76-ed715c297238" Dec 03 08:56:40 crc kubenswrapper[4573]: I1203 08:56:40.241334 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:40 crc kubenswrapper[4573]: I1203 08:56:40.241446 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:40 crc kubenswrapper[4573]: E1203 08:56:40.241658 4573 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 08:56:40 crc kubenswrapper[4573]: E1203 08:56:40.241736 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:44.241715808 +0000 UTC m=+1124.810095067 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "webhook-server-cert" not found Dec 03 08:56:40 crc kubenswrapper[4573]: E1203 08:56:40.241846 4573 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 08:56:40 crc kubenswrapper[4573]: E1203 08:56:40.241954 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:44.241932294 +0000 UTC m=+1124.810311723 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "metrics-server-cert" not found Dec 03 08:56:41 crc kubenswrapper[4573]: I1203 08:56:41.982535 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:41 crc kubenswrapper[4573]: E1203 08:56:41.982813 4573 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:41 crc kubenswrapper[4573]: E1203 08:56:41.982883 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert podName:22bca7e1-22a7-4ee7-852b-25fddeb8fce3 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:49.982860872 +0000 UTC m=+1130.551240131 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert") pod "infra-operator-controller-manager-57548d458d-gljjt" (UID: "22bca7e1-22a7-4ee7-852b-25fddeb8fce3") : secret "infra-operator-webhook-server-cert" not found Dec 03 08:56:43 crc kubenswrapper[4573]: I1203 08:56:43.034963 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:43 crc kubenswrapper[4573]: E1203 08:56:43.035546 4573 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:43 crc kubenswrapper[4573]: E1203 08:56:43.035637 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert podName:a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629 nodeName:}" failed. No retries permitted until 2025-12-03 08:56:51.035609945 +0000 UTC m=+1131.603989204 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" (UID: "a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 08:56:44 crc kubenswrapper[4573]: I1203 08:56:44.282117 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:44 crc kubenswrapper[4573]: I1203 08:56:44.282738 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:44 crc kubenswrapper[4573]: E1203 08:56:44.282591 4573 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 08:56:44 crc kubenswrapper[4573]: E1203 08:56:44.283022 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:52.2830008 +0000 UTC m=+1132.851380059 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "webhook-server-cert" not found Dec 03 08:56:44 crc kubenswrapper[4573]: E1203 08:56:44.282939 4573 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 08:56:44 crc kubenswrapper[4573]: E1203 08:56:44.283528 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs podName:5acea32c-fc74-40d8-b363-c64470ee880e nodeName:}" failed. No retries permitted until 2025-12-03 08:56:52.283516234 +0000 UTC m=+1132.851895493 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs") pod "openstack-operator-controller-manager-58f6d66c48-gwplc" (UID: "5acea32c-fc74-40d8-b363-c64470ee880e") : secret "metrics-server-cert" not found Dec 03 08:56:49 crc kubenswrapper[4573]: I1203 08:56:49.990233 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:50 crc kubenswrapper[4573]: I1203 08:56:50.024232 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22bca7e1-22a7-4ee7-852b-25fddeb8fce3-cert\") pod \"infra-operator-controller-manager-57548d458d-gljjt\" (UID: \"22bca7e1-22a7-4ee7-852b-25fddeb8fce3\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:50 crc kubenswrapper[4573]: I1203 08:56:50.175672 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:56:51 crc kubenswrapper[4573]: I1203 08:56:51.108425 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:51 crc kubenswrapper[4573]: I1203 08:56:51.114850 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg\" (UID: \"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:51 crc kubenswrapper[4573]: I1203 08:56:51.375326 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:56:52 crc kubenswrapper[4573]: I1203 08:56:52.328421 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:52 crc kubenswrapper[4573]: I1203 08:56:52.330310 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:52 crc kubenswrapper[4573]: I1203 08:56:52.340505 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-webhook-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:52 crc kubenswrapper[4573]: I1203 08:56:52.340505 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5acea32c-fc74-40d8-b363-c64470ee880e-metrics-certs\") pod \"openstack-operator-controller-manager-58f6d66c48-gwplc\" (UID: \"5acea32c-fc74-40d8-b363-c64470ee880e\") " pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:52 crc kubenswrapper[4573]: I1203 08:56:52.618361 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:56:52 crc kubenswrapper[4573]: E1203 08:56:52.972980 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 03 08:56:52 crc kubenswrapper[4573]: E1203 08:56:52.973357 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p5txz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-c5vc7_openstack-operators(e94079e6-8701-48bf-ab58-1867fdf6e46e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:56:53 crc kubenswrapper[4573]: E1203 08:56:53.549792 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 03 08:56:53 crc kubenswrapper[4573]: E1203 08:56:53.550176 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sgmfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-r6qvb_openstack-operators(9d6a307e-5cf2-4a23-921c-c5b562494cb0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:56:55 crc kubenswrapper[4573]: E1203 08:56:55.197945 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 03 08:56:55 crc kubenswrapper[4573]: E1203 08:56:55.198894 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pr8r5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-qphvx_openstack-operators(4613c040-3ca1-458c-8ff8-2b8858a7ad35): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:56:56 crc kubenswrapper[4573]: E1203 08:56:56.670194 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 03 08:56:56 crc kubenswrapper[4573]: E1203 08:56:56.670543 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7wprn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-2qnlp_openstack-operators(622de9fb-c4ba-4727-8c22-5fcb36f39751): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:56:57 crc kubenswrapper[4573]: E1203 08:56:57.292919 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 03 08:56:57 crc kubenswrapper[4573]: E1203 08:56:57.293334 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2c2vt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-jd6bv_openstack-operators(8f65995e-e70a-48e4-9504-8aa85d43ea9e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:02 crc kubenswrapper[4573]: E1203 08:57:02.150229 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 03 08:57:02 crc kubenswrapper[4573]: E1203 08:57:02.154965 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kknkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-hb6l4_openstack-operators(e0ae572b-e68c-48a0-8649-7eea884e61b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:02 crc kubenswrapper[4573]: E1203 08:57:02.918217 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 03 08:57:02 crc kubenswrapper[4573]: E1203 08:57:02.918962 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d44qp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-hl72w_openstack-operators(a2888fa7-1f19-48c8-853e-a3952cc93ec6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:03 crc kubenswrapper[4573]: E1203 08:57:03.629804 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 03 08:57:03 crc kubenswrapper[4573]: E1203 08:57:03.630653 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cgdl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-bnztb_openstack-operators(98fcc67c-249b-4700-8434-c8dd4e1646ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:04 crc kubenswrapper[4573]: E1203 08:57:04.298896 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 03 08:57:04 crc kubenswrapper[4573]: E1203 08:57:04.299221 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x97tw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-gnlgf_openstack-operators(ff1aaf12-2fcd-41d0-a575-a0e5447fc11f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:06 crc kubenswrapper[4573]: E1203 08:57:06.571469 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 03 08:57:06 crc kubenswrapper[4573]: E1203 08:57:06.571817 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v7cfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-wcbtn_openstack-operators(9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:10 crc kubenswrapper[4573]: E1203 08:57:10.224520 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 03 08:57:10 crc kubenswrapper[4573]: E1203 08:57:10.225297 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bh2km,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-l5vf4_openstack-operators(e6f9febf-6db7-4ea8-8b96-7a939102be2a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:11 crc kubenswrapper[4573]: E1203 08:57:11.150393 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 08:57:11 crc kubenswrapper[4573]: E1203 08:57:11.150680 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5d6pk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-sdrbz_openstack-operators(fd426f33-5f03-4fec-bde4-e72e9c762762): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:14 crc kubenswrapper[4573]: E1203 08:57:14.308759 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 08:57:14 crc kubenswrapper[4573]: E1203 08:57:14.310185 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p4lrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-hdhrp_openstack-operators(5aba2f13-92bd-4ea3-bfcb-7646909db04b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:14 crc kubenswrapper[4573]: E1203 08:57:14.904733 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 08:57:14 crc kubenswrapper[4573]: E1203 08:57:14.905092 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9gx5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-jbwlq_openstack-operators(b4436129-0e91-443b-8a97-61fa7f8cc2e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:14 crc kubenswrapper[4573]: E1203 08:57:14.906497 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" podUID="b4436129-0e91-443b-8a97-61fa7f8cc2e4" Dec 03 08:57:15 crc kubenswrapper[4573]: I1203 08:57:15.153868 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg"] Dec 03 08:57:15 crc kubenswrapper[4573]: W1203 08:57:15.278739 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0fafd12_f0eb_49a0_9d3a_7eefc2b5a629.slice/crio-ec00f380c12786218fe858c3379d39130eb02e6a21f6f501d5660339787b0847 WatchSource:0}: Error finding container ec00f380c12786218fe858c3379d39130eb02e6a21f6f501d5660339787b0847: Status 404 returned error can't find the container with id ec00f380c12786218fe858c3379d39130eb02e6a21f6f501d5660339787b0847 Dec 03 08:57:15 crc kubenswrapper[4573]: I1203 08:57:15.594371 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-gljjt"] Dec 03 08:57:15 crc kubenswrapper[4573]: I1203 08:57:15.627279 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc"] Dec 03 08:57:15 crc kubenswrapper[4573]: I1203 08:57:15.667039 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" event={"ID":"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629","Type":"ContainerStarted","Data":"ec00f380c12786218fe858c3379d39130eb02e6a21f6f501d5660339787b0847"} Dec 03 08:57:15 crc kubenswrapper[4573]: W1203 08:57:15.751989 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5acea32c_fc74_40d8_b363_c64470ee880e.slice/crio-da514425fe078d9b6b0cedaf2e1fe7be264c5e8455f06f3fe8c7d59f041b7ce5 WatchSource:0}: Error finding container da514425fe078d9b6b0cedaf2e1fe7be264c5e8455f06f3fe8c7d59f041b7ce5: Status 404 returned error can't find the container with id da514425fe078d9b6b0cedaf2e1fe7be264c5e8455f06f3fe8c7d59f041b7ce5 Dec 03 08:57:15 crc kubenswrapper[4573]: W1203 08:57:15.756352 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22bca7e1_22a7_4ee7_852b_25fddeb8fce3.slice/crio-5a3dd0506529b10fd48ff1ea26b07ddf33bfd677ede35278f5c49c6d851aea59 WatchSource:0}: Error finding container 5a3dd0506529b10fd48ff1ea26b07ddf33bfd677ede35278f5c49c6d851aea59: Status 404 returned error can't find the container with id 5a3dd0506529b10fd48ff1ea26b07ddf33bfd677ede35278f5c49c6d851aea59 Dec 03 08:57:16 crc kubenswrapper[4573]: I1203 08:57:16.673569 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" event={"ID":"5acea32c-fc74-40d8-b363-c64470ee880e","Type":"ContainerStarted","Data":"da514425fe078d9b6b0cedaf2e1fe7be264c5e8455f06f3fe8c7d59f041b7ce5"} Dec 03 08:57:16 crc kubenswrapper[4573]: I1203 08:57:16.674463 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" event={"ID":"22bca7e1-22a7-4ee7-852b-25fddeb8fce3","Type":"ContainerStarted","Data":"5a3dd0506529b10fd48ff1ea26b07ddf33bfd677ede35278f5c49c6d851aea59"} Dec 03 08:57:16 crc kubenswrapper[4573]: I1203 08:57:16.675651 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" event={"ID":"a855d891-6a41-477d-98d7-fcbba59aee28","Type":"ContainerStarted","Data":"e1dc8e3c43afbb8529431c59b7211ff8c150b5cf4765031beb3e953fd7c3da29"} Dec 03 08:57:16 crc kubenswrapper[4573]: I1203 08:57:16.676867 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" event={"ID":"fd5e6711-117e-42b2-90e6-cfddf410315d","Type":"ContainerStarted","Data":"5643a2ca56c863f6d1c34aaff9e1a0d547d9cd0da3630a1d2be9c0af2aa8ca48"} Dec 03 08:57:17 crc kubenswrapper[4573]: I1203 08:57:17.686616 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" event={"ID":"e27aea95-097f-4aa8-ae2b-7d212b8640b4","Type":"ContainerStarted","Data":"b6e781fde0007bf5f838baf84784ea2a1040c6cdc4cef48ba3caa9c237be8dd3"} Dec 03 08:57:17 crc kubenswrapper[4573]: I1203 08:57:17.688481 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" event={"ID":"ee80479f-97a6-43ef-a026-d5bf1931d962","Type":"ContainerStarted","Data":"6866674c66ae88f76b400307ec70094b3543f917737b48a37dcf4bd3c0df3e89"} Dec 03 08:57:17 crc kubenswrapper[4573]: I1203 08:57:17.690691 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" event={"ID":"6535c22e-f476-4067-b1ab-7e0ed5c23360","Type":"ContainerStarted","Data":"644e44595280e90ede8533e7d29c92b218bc9d42e7ea55d995e925bddb30e5e9"} Dec 03 08:57:20 crc kubenswrapper[4573]: I1203 08:57:20.753204 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" event={"ID":"5acea32c-fc74-40d8-b363-c64470ee880e","Type":"ContainerStarted","Data":"6dddc155ff34a914d97c9ef28bca2f37d5fbedca16af728c377c7f9720098940"} Dec 03 08:57:20 crc kubenswrapper[4573]: I1203 08:57:20.754606 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:57:20 crc kubenswrapper[4573]: I1203 08:57:20.808261 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" podStartSLOduration=44.808223824 podStartE2EDuration="44.808223824s" podCreationTimestamp="2025-12-03 08:56:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:57:20.79706474 +0000 UTC m=+1161.365444019" watchObservedRunningTime="2025-12-03 08:57:20.808223824 +0000 UTC m=+1161.376603083" Dec 03 08:57:21 crc kubenswrapper[4573]: I1203 08:57:21.765518 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" event={"ID":"57f767e9-7c3a-4707-9b76-ed715c297238","Type":"ContainerStarted","Data":"dc91fba745179a0759b8b107c96c70a39d1cf7e47ef1753c8b6beeab9818523f"} Dec 03 08:57:22 crc kubenswrapper[4573]: E1203 08:57:22.333323 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 08:57:22 crc kubenswrapper[4573]: E1203 08:57:22.333560 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sgmfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-r6qvb_openstack-operators(9d6a307e-5cf2-4a23-921c-c5b562494cb0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:22 crc kubenswrapper[4573]: E1203 08:57:22.335030 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" podUID="9d6a307e-5cf2-4a23-921c-c5b562494cb0" Dec 03 08:57:22 crc kubenswrapper[4573]: E1203 08:57:22.515065 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 08:57:22 crc kubenswrapper[4573]: E1203 08:57:22.515374 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pr8r5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-qphvx_openstack-operators(4613c040-3ca1-458c-8ff8-2b8858a7ad35): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:57:22 crc kubenswrapper[4573]: E1203 08:57:22.516761 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" podUID="4613c040-3ca1-458c-8ff8-2b8858a7ad35" Dec 03 08:57:22 crc kubenswrapper[4573]: I1203 08:57:22.775182 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 08:57:23 crc kubenswrapper[4573]: E1203 08:57:23.904040 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" podUID="98fcc67c-249b-4700-8434-c8dd4e1646ae" Dec 03 08:57:23 crc kubenswrapper[4573]: E1203 08:57:23.982824 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" podUID="622de9fb-c4ba-4727-8c22-5fcb36f39751" Dec 03 08:57:23 crc kubenswrapper[4573]: E1203 08:57:23.992495 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" podUID="e94079e6-8701-48bf-ab58-1867fdf6e46e" Dec 03 08:57:24 crc kubenswrapper[4573]: E1203 08:57:24.644709 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" podUID="8f65995e-e70a-48e4-9504-8aa85d43ea9e" Dec 03 08:57:24 crc kubenswrapper[4573]: E1203 08:57:24.705312 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" podUID="e6f9febf-6db7-4ea8-8b96-7a939102be2a" Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.841493 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" event={"ID":"9d6a307e-5cf2-4a23-921c-c5b562494cb0","Type":"ContainerStarted","Data":"281fe6e6f6ce3e55b30b905b72130e70bdad34a97c151549d2e4f7ae223ad0b4"} Dec 03 08:57:24 crc kubenswrapper[4573]: E1203 08:57:24.857109 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" podUID="ff1aaf12-2fcd-41d0-a575-a0e5447fc11f" Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.857649 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" event={"ID":"4613c040-3ca1-458c-8ff8-2b8858a7ad35","Type":"ContainerStarted","Data":"90b1ef753cd6b079edc573daea945f793405e96ef514fbe3ccf40becd81566b9"} Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.870497 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" event={"ID":"57f767e9-7c3a-4707-9b76-ed715c297238","Type":"ContainerStarted","Data":"88f1bd4f6fee65b1a27e087945a7d491d08dc0671d72c2c2974df8243dc3b072"} Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.871795 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.885709 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" event={"ID":"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629","Type":"ContainerStarted","Data":"87e8ffdb4f2763363b75f39863aa26bbfc6ac37ccd8770d19774c5abdf7e3f95"} Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.892528 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" event={"ID":"98fcc67c-249b-4700-8434-c8dd4e1646ae","Type":"ContainerStarted","Data":"932e907ed7e7c05ba848073381752862afe9bcae92427f9cadb1fae550637986"} Dec 03 08:57:24 crc kubenswrapper[4573]: E1203 08:57:24.917660 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" podUID="fd426f33-5f03-4fec-bde4-e72e9c762762" Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.927748 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" event={"ID":"e6f9febf-6db7-4ea8-8b96-7a939102be2a","Type":"ContainerStarted","Data":"d1e97dc7a5528f95836afb489d4a919993fbd24abdf1377074602a9de9e0d23b"} Dec 03 08:57:24 crc kubenswrapper[4573]: E1203 08:57:24.931712 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" podUID="e6f9febf-6db7-4ea8-8b96-7a939102be2a" Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.949570 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" podStartSLOduration=13.290767928 podStartE2EDuration="49.94953618s" podCreationTimestamp="2025-12-03 08:56:35 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.666430963 +0000 UTC m=+1119.234810222" lastFinishedPulling="2025-12-03 08:57:15.325199215 +0000 UTC m=+1155.893578474" observedRunningTime="2025-12-03 08:57:24.946680322 +0000 UTC m=+1165.515059581" watchObservedRunningTime="2025-12-03 08:57:24.94953618 +0000 UTC m=+1165.517915439" Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.973498 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" event={"ID":"6535c22e-f476-4067-b1ab-7e0ed5c23360","Type":"ContainerStarted","Data":"b76bd5419c07df2ad4b33428b48fb9378a9d7d2453fecaf4d2ad82d5bc9d57d7"} Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.974824 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.985946 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" Dec 03 08:57:24 crc kubenswrapper[4573]: E1203 08:57:24.991836 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" podUID="a2888fa7-1f19-48c8-853e-a3952cc93ec6" Dec 03 08:57:24 crc kubenswrapper[4573]: I1203 08:57:24.992742 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" event={"ID":"622de9fb-c4ba-4727-8c22-5fcb36f39751","Type":"ContainerStarted","Data":"4cf886c41653dc0f9603e9fa1affc381c2f13deb63ed813e385521ad57ab08d8"} Dec 03 08:57:25 crc kubenswrapper[4573]: E1203 08:57:25.013654 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" podUID="e0ae572b-e68c-48a0-8649-7eea884e61b0" Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.042008 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" event={"ID":"22bca7e1-22a7-4ee7-852b-25fddeb8fce3","Type":"ContainerStarted","Data":"b45a4047f0015459164e2cfc28b9c040916482ae92133ea1ac5dd2d93177830e"} Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.056152 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" event={"ID":"e94079e6-8701-48bf-ab58-1867fdf6e46e","Type":"ContainerStarted","Data":"1410db91849d56556934ba4aebde1e3693dd00992fedca033f0d62e28dd9bd59"} Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.084388 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" event={"ID":"e27aea95-097f-4aa8-ae2b-7d212b8640b4","Type":"ContainerStarted","Data":"d4d0a1c93e3ffb6d607354ca3b80199544301d8ac2badb6a44fc6d3ee193d69e"} Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.085378 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.091308 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" Dec 03 08:57:25 crc kubenswrapper[4573]: E1203 08:57:25.114836 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" podUID="5aba2f13-92bd-4ea3-bfcb-7646909db04b" Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.115151 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" event={"ID":"8f65995e-e70a-48e4-9504-8aa85d43ea9e","Type":"ContainerStarted","Data":"3d556243fb221c366c5eae09e3feb1a81e563daf1f3fa1cc26ea0028344a3004"} Dec 03 08:57:25 crc kubenswrapper[4573]: E1203 08:57:25.132532 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" podUID="9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b" Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.146401 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" event={"ID":"ee80479f-97a6-43ef-a026-d5bf1931d962","Type":"ContainerStarted","Data":"9c6f238a5b681c4b03d9c4c5d8adc587ec373e005ce6540ff6060cdaf38f48ac"} Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.147969 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.156930 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.179213 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wdvlj" podStartSLOduration=5.457955778 podStartE2EDuration="50.179174274s" podCreationTimestamp="2025-12-03 08:56:35 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.558893565 +0000 UTC m=+1119.127272824" lastFinishedPulling="2025-12-03 08:57:23.280112071 +0000 UTC m=+1163.848491320" observedRunningTime="2025-12-03 08:57:25.175758471 +0000 UTC m=+1165.744137740" watchObservedRunningTime="2025-12-03 08:57:25.179174274 +0000 UTC m=+1165.747553533" Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.236359 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-nx7mq" podStartSLOduration=5.264709612 podStartE2EDuration="52.236335271s" podCreationTimestamp="2025-12-03 08:56:33 +0000 UTC" firstStartedPulling="2025-12-03 08:56:36.525518982 +0000 UTC m=+1117.093898241" lastFinishedPulling="2025-12-03 08:57:23.497144641 +0000 UTC m=+1164.065523900" observedRunningTime="2025-12-03 08:57:25.235076547 +0000 UTC m=+1165.803455826" watchObservedRunningTime="2025-12-03 08:57:25.236335271 +0000 UTC m=+1165.804714530" Dec 03 08:57:25 crc kubenswrapper[4573]: I1203 08:57:25.304923 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-vl9xb" podStartSLOduration=5.340559154 podStartE2EDuration="51.304894358s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.287313861 +0000 UTC m=+1117.855693120" lastFinishedPulling="2025-12-03 08:57:23.251649065 +0000 UTC m=+1163.820028324" observedRunningTime="2025-12-03 08:57:25.287095153 +0000 UTC m=+1165.855474412" watchObservedRunningTime="2025-12-03 08:57:25.304894358 +0000 UTC m=+1165.873273607" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.082023 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-nbfxj" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.198599 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" event={"ID":"9d6a307e-5cf2-4a23-921c-c5b562494cb0","Type":"ContainerStarted","Data":"d730b1ffcc966d1b2b008184dbd82cf325f09bc208d7fb9540a38c46f0843f78"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.205898 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" event={"ID":"fd426f33-5f03-4fec-bde4-e72e9c762762","Type":"ContainerStarted","Data":"e7012b91354ff0c5d46b5678d41ee29ab7a8d458d4e85c140e926e5a80826bf7"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.220398 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" event={"ID":"e0ae572b-e68c-48a0-8649-7eea884e61b0","Type":"ContainerStarted","Data":"0696efec4cb41e7c5861c1fd0fe7c1b5b3dad4db0f1ccf74f158ada77c35e5f4"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.230650 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" event={"ID":"9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b","Type":"ContainerStarted","Data":"55989127ed573430a9178b4581a33b7a24d395f5529ebc7b9b49c3e04d715749"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.237974 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" event={"ID":"a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629","Type":"ContainerStarted","Data":"30c5ca55abe07b413b8d64a54ee459b9ecffb560d32fb0771494c93b9216810a"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.238671 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.272608 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" event={"ID":"5aba2f13-92bd-4ea3-bfcb-7646909db04b","Type":"ContainerStarted","Data":"43d3694db5db6deeaf08129ed600489f4e18dcb768bef3e71882c33a0b90a7a2"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.297725 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" event={"ID":"ff1aaf12-2fcd-41d0-a575-a0e5447fc11f","Type":"ContainerStarted","Data":"846b2bb059b4fc37ed0910d3f11189d8389740458cb1871e63c693d9ee48eddd"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.312616 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" event={"ID":"a2888fa7-1f19-48c8-853e-a3952cc93ec6","Type":"ContainerStarted","Data":"2773973c4c29b70eb15188d49ebc189f949f08f7b98c686d16d58a897bde9b4c"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.430839 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" podStartSLOduration=44.480664798 podStartE2EDuration="52.430813424s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:57:15.291025874 +0000 UTC m=+1155.859405143" lastFinishedPulling="2025-12-03 08:57:23.24117451 +0000 UTC m=+1163.809553769" observedRunningTime="2025-12-03 08:57:26.427539695 +0000 UTC m=+1166.995918954" watchObservedRunningTime="2025-12-03 08:57:26.430813424 +0000 UTC m=+1166.999192683" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.434563 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" event={"ID":"a855d891-6a41-477d-98d7-fcbba59aee28","Type":"ContainerStarted","Data":"da3a8969d38e045784e7202c10ed6fa3b998d38d930d9b6d75af73f5ccff774d"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.440315 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.456620 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.477705 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" event={"ID":"fd5e6711-117e-42b2-90e6-cfddf410315d","Type":"ContainerStarted","Data":"2117ec013f6372b8a72c7ebb2cd9ec6ef131f99ba6d7df5af10879796dd08d79"} Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.478097 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.481021 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.537974 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-s8679" podStartSLOduration=6.8424080499999995 podStartE2EDuration="52.537949363s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.79725477 +0000 UTC m=+1118.365634029" lastFinishedPulling="2025-12-03 08:57:23.492796083 +0000 UTC m=+1164.061175342" observedRunningTime="2025-12-03 08:57:26.532156165 +0000 UTC m=+1167.100535444" watchObservedRunningTime="2025-12-03 08:57:26.537949363 +0000 UTC m=+1167.106328612" Dec 03 08:57:26 crc kubenswrapper[4573]: I1203 08:57:26.733178 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-d77h9" podStartSLOduration=7.66742423 podStartE2EDuration="52.733142119s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.496473324 +0000 UTC m=+1119.064852583" lastFinishedPulling="2025-12-03 08:57:23.562191203 +0000 UTC m=+1164.130570472" observedRunningTime="2025-12-03 08:57:26.720805593 +0000 UTC m=+1167.289184842" watchObservedRunningTime="2025-12-03 08:57:26.733142119 +0000 UTC m=+1167.301521378" Dec 03 08:57:27 crc kubenswrapper[4573]: E1203 08:57:27.106147 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" podUID="b4436129-0e91-443b-8a97-61fa7f8cc2e4" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.509016 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" event={"ID":"8f65995e-e70a-48e4-9504-8aa85d43ea9e","Type":"ContainerStarted","Data":"584a4f17df44cf4887b4e1167d497e8a770b31094a895df935b04d12da009a2a"} Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.526938 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" event={"ID":"98fcc67c-249b-4700-8434-c8dd4e1646ae","Type":"ContainerStarted","Data":"7bcb83be75f618578ad02c83a6042ded7682d1ef37e8ff2b1368428be4cba2e4"} Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.528254 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.533262 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" event={"ID":"4613c040-3ca1-458c-8ff8-2b8858a7ad35","Type":"ContainerStarted","Data":"ed4ada96252ba840475c17a91b3688193391c84837a59e3abcd2f8fa728d0b27"} Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.534633 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.630350 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" event={"ID":"622de9fb-c4ba-4727-8c22-5fcb36f39751","Type":"ContainerStarted","Data":"f541d83ccc0f34084be9707f6d0b6e7ffab75059b08e4f152d8c8ea480b73569"} Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.630511 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.637288 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" event={"ID":"22bca7e1-22a7-4ee7-852b-25fddeb8fce3","Type":"ContainerStarted","Data":"daf24f8cf2cac73f68812db1545666a94bb5274d1022282456b54fbc815a72d5"} Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.640387 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.660701 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" podStartSLOduration=4.583163612 podStartE2EDuration="52.660679512s" podCreationTimestamp="2025-12-03 08:56:35 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.389186252 +0000 UTC m=+1118.957565511" lastFinishedPulling="2025-12-03 08:57:26.466702162 +0000 UTC m=+1167.035081411" observedRunningTime="2025-12-03 08:57:27.621855065 +0000 UTC m=+1168.190234334" watchObservedRunningTime="2025-12-03 08:57:27.660679512 +0000 UTC m=+1168.229058771" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.669589 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" event={"ID":"e94079e6-8701-48bf-ab58-1867fdf6e46e","Type":"ContainerStarted","Data":"c35ddff45e5d2f8183ff06bdf06d429742d7f31a8118ad676091382850df6be8"} Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.670567 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.744090 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" podStartSLOduration=5.178162351 podStartE2EDuration="53.744072293s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.702312704 +0000 UTC m=+1118.270691963" lastFinishedPulling="2025-12-03 08:57:26.268222646 +0000 UTC m=+1166.836601905" observedRunningTime="2025-12-03 08:57:27.729954679 +0000 UTC m=+1168.298333938" watchObservedRunningTime="2025-12-03 08:57:27.744072293 +0000 UTC m=+1168.312451552" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.813388 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" podStartSLOduration=8.402360358 podStartE2EDuration="53.813359061s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.145658449 +0000 UTC m=+1118.714037708" lastFinishedPulling="2025-12-03 08:57:23.556657152 +0000 UTC m=+1164.125036411" observedRunningTime="2025-12-03 08:57:27.81332714 +0000 UTC m=+1168.381706399" watchObservedRunningTime="2025-12-03 08:57:27.813359061 +0000 UTC m=+1168.381738320" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.887258 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" podStartSLOduration=8.819301213 podStartE2EDuration="53.887230382s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.429014477 +0000 UTC m=+1118.997393736" lastFinishedPulling="2025-12-03 08:57:23.496943646 +0000 UTC m=+1164.065322905" observedRunningTime="2025-12-03 08:57:27.858016277 +0000 UTC m=+1168.426395546" watchObservedRunningTime="2025-12-03 08:57:27.887230382 +0000 UTC m=+1168.455609631" Dec 03 08:57:27 crc kubenswrapper[4573]: I1203 08:57:27.955593 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" podStartSLOduration=4.413287465 podStartE2EDuration="52.955559344s" podCreationTimestamp="2025-12-03 08:56:35 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.104117947 +0000 UTC m=+1118.672497206" lastFinishedPulling="2025-12-03 08:57:26.646389826 +0000 UTC m=+1167.214769085" observedRunningTime="2025-12-03 08:57:27.945571132 +0000 UTC m=+1168.513950391" watchObservedRunningTime="2025-12-03 08:57:27.955559344 +0000 UTC m=+1168.523938603" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.028980 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" podStartSLOduration=4.6655614 podStartE2EDuration="54.028956583s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.321461871 +0000 UTC m=+1117.889841130" lastFinishedPulling="2025-12-03 08:57:26.684857044 +0000 UTC m=+1167.253236313" observedRunningTime="2025-12-03 08:57:28.009617116 +0000 UTC m=+1168.577996375" watchObservedRunningTime="2025-12-03 08:57:28.028956583 +0000 UTC m=+1168.597335842" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.068168 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" podStartSLOduration=46.554144782 podStartE2EDuration="54.06814222s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:57:15.76049903 +0000 UTC m=+1156.328878289" lastFinishedPulling="2025-12-03 08:57:23.274496458 +0000 UTC m=+1163.842875727" observedRunningTime="2025-12-03 08:57:28.061586611 +0000 UTC m=+1168.629965880" watchObservedRunningTime="2025-12-03 08:57:28.06814222 +0000 UTC m=+1168.636521469" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.672810 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" event={"ID":"ff1aaf12-2fcd-41d0-a575-a0e5447fc11f","Type":"ContainerStarted","Data":"2fb47c4a87772afa67edc693d6ed3cf535bbac60967e53cca285de35e6d630f7"} Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.674646 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.679128 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" event={"ID":"5aba2f13-92bd-4ea3-bfcb-7646909db04b","Type":"ContainerStarted","Data":"341d660ee2bda80f9f3f69a1dd7440b6525d0145581b5790fb3d21edc0d5b79b"} Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.679677 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.685173 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" event={"ID":"fd426f33-5f03-4fec-bde4-e72e9c762762","Type":"ContainerStarted","Data":"f9e122d3a2ffa5a55c180968eea5fff338e9ab730921456f1e41dece28e13fa5"} Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.685395 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.688862 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" event={"ID":"e0ae572b-e68c-48a0-8649-7eea884e61b0","Type":"ContainerStarted","Data":"54c2547b07af70cc7b8f7c6adf3046d41690c97e73481c17bce65932feae2046"} Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.689076 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.692586 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" event={"ID":"a2888fa7-1f19-48c8-853e-a3952cc93ec6","Type":"ContainerStarted","Data":"2d279ca5ce2a58688c95cc39aa051c9cd9691f94b9d46354514a566911eed5ed"} Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.692807 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.700086 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" event={"ID":"9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b","Type":"ContainerStarted","Data":"7a3370bb06fd656c197f342e39152377f3658d495a113407325f3cd728ee7357"} Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.704537 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.704592 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.962155 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" podStartSLOduration=5.249590007 podStartE2EDuration="54.962123419s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.473906363 +0000 UTC m=+1118.042285622" lastFinishedPulling="2025-12-03 08:57:27.186439765 +0000 UTC m=+1167.754819034" observedRunningTime="2025-12-03 08:57:28.70109948 +0000 UTC m=+1169.269478739" watchObservedRunningTime="2025-12-03 08:57:28.962123419 +0000 UTC m=+1169.530502678" Dec 03 08:57:28 crc kubenswrapper[4573]: I1203 08:57:28.993768 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" podStartSLOduration=5.737867925 podStartE2EDuration="54.99374423s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.932605576 +0000 UTC m=+1118.500984835" lastFinishedPulling="2025-12-03 08:57:27.188481861 +0000 UTC m=+1167.756861140" observedRunningTime="2025-12-03 08:57:28.991942542 +0000 UTC m=+1169.560321801" watchObservedRunningTime="2025-12-03 08:57:28.99374423 +0000 UTC m=+1169.562123489" Dec 03 08:57:29 crc kubenswrapper[4573]: I1203 08:57:29.087192 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" Dec 03 08:57:29 crc kubenswrapper[4573]: I1203 08:57:29.088251 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" podStartSLOduration=5.757292005 podStartE2EDuration="55.088211734s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.820285467 +0000 UTC m=+1118.388664716" lastFinishedPulling="2025-12-03 08:57:27.151205196 +0000 UTC m=+1167.719584445" observedRunningTime="2025-12-03 08:57:29.069949356 +0000 UTC m=+1169.638328615" watchObservedRunningTime="2025-12-03 08:57:29.088211734 +0000 UTC m=+1169.656590993" Dec 03 08:57:29 crc kubenswrapper[4573]: I1203 08:57:29.153385 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" podStartSLOduration=6.409782801 podStartE2EDuration="56.153355638s" podCreationTimestamp="2025-12-03 08:56:33 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.407505185 +0000 UTC m=+1117.975884444" lastFinishedPulling="2025-12-03 08:57:27.151078022 +0000 UTC m=+1167.719457281" observedRunningTime="2025-12-03 08:57:29.137711182 +0000 UTC m=+1169.706090441" watchObservedRunningTime="2025-12-03 08:57:29.153355638 +0000 UTC m=+1169.721734897" Dec 03 08:57:29 crc kubenswrapper[4573]: I1203 08:57:29.224870 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" podStartSLOduration=5.815482801 podStartE2EDuration="55.224844945s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:37.824359429 +0000 UTC m=+1118.392738688" lastFinishedPulling="2025-12-03 08:57:27.233721573 +0000 UTC m=+1167.802100832" observedRunningTime="2025-12-03 08:57:29.218741699 +0000 UTC m=+1169.787120958" watchObservedRunningTime="2025-12-03 08:57:29.224844945 +0000 UTC m=+1169.793224204" Dec 03 08:57:29 crc kubenswrapper[4573]: I1203 08:57:29.227205 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" podStartSLOduration=6.348205829 podStartE2EDuration="55.227193189s" podCreationTimestamp="2025-12-03 08:56:34 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.269697967 +0000 UTC m=+1118.838077226" lastFinishedPulling="2025-12-03 08:57:27.148685327 +0000 UTC m=+1167.717064586" observedRunningTime="2025-12-03 08:57:29.193761029 +0000 UTC m=+1169.762140288" watchObservedRunningTime="2025-12-03 08:57:29.227193189 +0000 UTC m=+1169.795572448" Dec 03 08:57:29 crc kubenswrapper[4573]: I1203 08:57:29.706511 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" Dec 03 08:57:29 crc kubenswrapper[4573]: I1203 08:57:29.711332 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r6qvb" Dec 03 08:57:30 crc kubenswrapper[4573]: I1203 08:57:30.182659 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-gljjt" Dec 03 08:57:32 crc kubenswrapper[4573]: I1203 08:57:32.625368 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-58f6d66c48-gwplc" Dec 03 08:57:34 crc kubenswrapper[4573]: I1203 08:57:34.317803 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-hb6l4" Dec 03 08:57:34 crc kubenswrapper[4573]: I1203 08:57:34.469423 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2qnlp" Dec 03 08:57:34 crc kubenswrapper[4573]: I1203 08:57:34.513034 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gnlgf" Dec 03 08:57:34 crc kubenswrapper[4573]: I1203 08:57:34.762509 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-wcbtn" Dec 03 08:57:34 crc kubenswrapper[4573]: I1203 08:57:34.821076 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-hdhrp" Dec 03 08:57:34 crc kubenswrapper[4573]: I1203 08:57:34.959416 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hl72w" Dec 03 08:57:35 crc kubenswrapper[4573]: I1203 08:57:35.139917 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-qphvx" Dec 03 08:57:35 crc kubenswrapper[4573]: I1203 08:57:35.374574 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-bnztb" Dec 03 08:57:35 crc kubenswrapper[4573]: I1203 08:57:35.411759 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-sdrbz" Dec 03 08:57:35 crc kubenswrapper[4573]: I1203 08:57:35.826698 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c5vc7" Dec 03 08:57:36 crc kubenswrapper[4573]: I1203 08:57:36.182756 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-jd6bv" Dec 03 08:57:38 crc kubenswrapper[4573]: I1203 08:57:38.791025 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" event={"ID":"e6f9febf-6db7-4ea8-8b96-7a939102be2a","Type":"ContainerStarted","Data":"734881bccb82c01e68702e3bee6327dce3624eb18e1c035b86bd9137d1b0518e"} Dec 03 08:57:38 crc kubenswrapper[4573]: I1203 08:57:38.792314 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" Dec 03 08:57:38 crc kubenswrapper[4573]: I1203 08:57:38.823092 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" podStartSLOduration=3.9718336819999998 podStartE2EDuration="1m3.823066589s" podCreationTimestamp="2025-12-03 08:56:35 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.588618814 +0000 UTC m=+1119.156998073" lastFinishedPulling="2025-12-03 08:57:38.439851721 +0000 UTC m=+1179.008230980" observedRunningTime="2025-12-03 08:57:38.816930362 +0000 UTC m=+1179.385309631" watchObservedRunningTime="2025-12-03 08:57:38.823066589 +0000 UTC m=+1179.391445848" Dec 03 08:57:39 crc kubenswrapper[4573]: I1203 08:57:39.803294 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" event={"ID":"b4436129-0e91-443b-8a97-61fa7f8cc2e4","Type":"ContainerStarted","Data":"50fadcc90e017d015063b5428a20df9ee1195eaab593259e113893031b39d05b"} Dec 03 08:57:39 crc kubenswrapper[4573]: I1203 08:57:39.827531 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-jbwlq" podStartSLOduration=2.899286266 podStartE2EDuration="1m3.827501057s" podCreationTimestamp="2025-12-03 08:56:36 +0000 UTC" firstStartedPulling="2025-12-03 08:56:38.599565432 +0000 UTC m=+1119.167944681" lastFinishedPulling="2025-12-03 08:57:39.527780213 +0000 UTC m=+1180.096159472" observedRunningTime="2025-12-03 08:57:39.824757432 +0000 UTC m=+1180.393136701" watchObservedRunningTime="2025-12-03 08:57:39.827501057 +0000 UTC m=+1180.395880316" Dec 03 08:57:46 crc kubenswrapper[4573]: I1203 08:57:46.557948 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-l5vf4" Dec 03 08:57:56 crc kubenswrapper[4573]: I1203 08:57:56.942864 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:57:56 crc kubenswrapper[4573]: I1203 08:57:56.943354 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.241494 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7ncg"] Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.244234 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.247424 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-jv8jk" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.248664 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.250457 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.254306 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.266681 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7ncg"] Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.331555 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6p7tr"] Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.333333 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.336275 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.403903 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6p7tr"] Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.418432 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/48ae045b-472f-4957-83ba-5020dde8cde0-kube-api-access-nhqw9\") pod \"dnsmasq-dns-675f4bcbfc-h7ncg\" (UID: \"48ae045b-472f-4957-83ba-5020dde8cde0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.418503 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48ae045b-472f-4957-83ba-5020dde8cde0-config\") pod \"dnsmasq-dns-675f4bcbfc-h7ncg\" (UID: \"48ae045b-472f-4957-83ba-5020dde8cde0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.519854 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-config\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.519921 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.519979 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/48ae045b-472f-4957-83ba-5020dde8cde0-kube-api-access-nhqw9\") pod \"dnsmasq-dns-675f4bcbfc-h7ncg\" (UID: \"48ae045b-472f-4957-83ba-5020dde8cde0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.520024 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48ae045b-472f-4957-83ba-5020dde8cde0-config\") pod \"dnsmasq-dns-675f4bcbfc-h7ncg\" (UID: \"48ae045b-472f-4957-83ba-5020dde8cde0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.520078 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhdch\" (UniqueName: \"kubernetes.io/projected/1010ea0b-292b-4e97-b46d-e3a9794530b0-kube-api-access-fhdch\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.521287 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48ae045b-472f-4957-83ba-5020dde8cde0-config\") pod \"dnsmasq-dns-675f4bcbfc-h7ncg\" (UID: \"48ae045b-472f-4957-83ba-5020dde8cde0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.548536 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/48ae045b-472f-4957-83ba-5020dde8cde0-kube-api-access-nhqw9\") pod \"dnsmasq-dns-675f4bcbfc-h7ncg\" (UID: \"48ae045b-472f-4957-83ba-5020dde8cde0\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.563276 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.621177 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhdch\" (UniqueName: \"kubernetes.io/projected/1010ea0b-292b-4e97-b46d-e3a9794530b0-kube-api-access-fhdch\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.621835 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-config\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.621876 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.623126 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.623834 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-config\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.649115 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhdch\" (UniqueName: \"kubernetes.io/projected/1010ea0b-292b-4e97-b46d-e3a9794530b0-kube-api-access-fhdch\") pod \"dnsmasq-dns-78dd6ddcc-6p7tr\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.654942 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:13 crc kubenswrapper[4573]: I1203 08:58:13.897845 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7ncg"] Dec 03 08:58:14 crc kubenswrapper[4573]: I1203 08:58:14.125564 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" event={"ID":"48ae045b-472f-4957-83ba-5020dde8cde0","Type":"ContainerStarted","Data":"12abb32a81bb0958a339b63e6e53a43643dce4287cc082c53ad575a992774817"} Dec 03 08:58:14 crc kubenswrapper[4573]: I1203 08:58:14.225992 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6p7tr"] Dec 03 08:58:14 crc kubenswrapper[4573]: W1203 08:58:14.226091 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1010ea0b_292b_4e97_b46d_e3a9794530b0.slice/crio-419f5ed0dc3ef971d4e612fb98a434a5ecb3e42c17de13811379628d2696acd7 WatchSource:0}: Error finding container 419f5ed0dc3ef971d4e612fb98a434a5ecb3e42c17de13811379628d2696acd7: Status 404 returned error can't find the container with id 419f5ed0dc3ef971d4e612fb98a434a5ecb3e42c17de13811379628d2696acd7 Dec 03 08:58:15 crc kubenswrapper[4573]: I1203 08:58:15.146588 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" event={"ID":"1010ea0b-292b-4e97-b46d-e3a9794530b0","Type":"ContainerStarted","Data":"419f5ed0dc3ef971d4e612fb98a434a5ecb3e42c17de13811379628d2696acd7"} Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.420510 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7ncg"] Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.482951 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-h42zx"] Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.499868 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.521734 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-h42zx"] Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.684211 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-config\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.684426 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l29p\" (UniqueName: \"kubernetes.io/projected/12f89059-a1f2-45fa-be11-e4bdc6b6f936-kube-api-access-7l29p\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.684571 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-dns-svc\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.786466 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-config\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.786611 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l29p\" (UniqueName: \"kubernetes.io/projected/12f89059-a1f2-45fa-be11-e4bdc6b6f936-kube-api-access-7l29p\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.786694 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-dns-svc\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.787734 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-config\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.788009 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-dns-svc\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:16 crc kubenswrapper[4573]: I1203 08:58:16.835638 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l29p\" (UniqueName: \"kubernetes.io/projected/12f89059-a1f2-45fa-be11-e4bdc6b6f936-kube-api-access-7l29p\") pod \"dnsmasq-dns-666b6646f7-h42zx\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.073705 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6p7tr"] Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.265987 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.316909 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b7qxz"] Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.318710 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.372468 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b7qxz"] Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.481533 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-config\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.481655 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.481707 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjtl8\" (UniqueName: \"kubernetes.io/projected/4b728cb5-fb6d-48e7-b170-238f76b6ec51-kube-api-access-mjtl8\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.583943 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.584028 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjtl8\" (UniqueName: \"kubernetes.io/projected/4b728cb5-fb6d-48e7-b170-238f76b6ec51-kube-api-access-mjtl8\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.584088 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-config\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.585082 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-config\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.585724 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.633577 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjtl8\" (UniqueName: \"kubernetes.io/projected/4b728cb5-fb6d-48e7-b170-238f76b6ec51-kube-api-access-mjtl8\") pod \"dnsmasq-dns-57d769cc4f-b7qxz\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.676075 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.753738 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.758998 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.767316 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.769751 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.769809 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.770007 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.770094 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.770143 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.770341 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-fzd4g" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.770484 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.894606 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.894713 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.894896 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.894931 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-config-data\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.894952 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.894970 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.894997 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.895029 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kzkl\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-kube-api-access-5kzkl\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.895138 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7087e292-a9fd-4420-bb4a-eb2d50355082-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.895226 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7087e292-a9fd-4420-bb4a-eb2d50355082-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:17 crc kubenswrapper[4573]: I1203 08:58:17.895266 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997306 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7087e292-a9fd-4420-bb4a-eb2d50355082-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997379 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997418 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997440 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997498 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997522 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-config-data\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997543 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997561 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997580 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997604 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kzkl\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-kube-api-access-5kzkl\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.997643 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7087e292-a9fd-4420-bb4a-eb2d50355082-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:17.999169 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-config-data\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.000552 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.000634 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.001446 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.002134 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.002650 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.006795 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7087e292-a9fd-4420-bb4a-eb2d50355082-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.007784 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7087e292-a9fd-4420-bb4a-eb2d50355082-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.009127 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.017627 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.047028 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kzkl\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-kube-api-access-5kzkl\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.093796 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.160728 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.270337 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-h42zx"] Dec 03 08:58:18 crc kubenswrapper[4573]: W1203 08:58:18.292880 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12f89059_a1f2_45fa_be11_e4bdc6b6f936.slice/crio-01276902add22aa019d40c1a2c57cef64747879bcb25b304d2961074dd37f51f WatchSource:0}: Error finding container 01276902add22aa019d40c1a2c57cef64747879bcb25b304d2961074dd37f51f: Status 404 returned error can't find the container with id 01276902add22aa019d40c1a2c57cef64747879bcb25b304d2961074dd37f51f Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.349283 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" event={"ID":"12f89059-a1f2-45fa-be11-e4bdc6b6f936","Type":"ContainerStarted","Data":"01276902add22aa019d40c1a2c57cef64747879bcb25b304d2961074dd37f51f"} Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.402117 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b7qxz"] Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.431341 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.443926 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.452335 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.457597 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.457967 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.458250 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.459295 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hb2p4" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.459527 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.459761 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.497374 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.541175 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9988868b-f38f-4e95-af61-1834e0e67a8b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.543910 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.544171 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.549299 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9988868b-f38f-4e95-af61-1834e0e67a8b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.549412 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.549443 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.549501 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.549793 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.549853 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6tm9\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-kube-api-access-h6tm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.549981 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.550133 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652425 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652497 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652522 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6tm9\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-kube-api-access-h6tm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652549 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652582 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652631 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9988868b-f38f-4e95-af61-1834e0e67a8b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652666 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652698 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652725 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9988868b-f38f-4e95-af61-1834e0e67a8b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652745 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.652769 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.653356 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.653471 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.654805 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.655293 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.655622 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.656200 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.665756 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9988868b-f38f-4e95-af61-1834e0e67a8b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.666793 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.667718 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9988868b-f38f-4e95-af61-1834e0e67a8b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.670255 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.674901 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6tm9\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-kube-api-access-h6tm9\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.679156 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.792864 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:58:18 crc kubenswrapper[4573]: I1203 08:58:18.974472 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 08:58:18 crc kubenswrapper[4573]: W1203 08:58:18.992649 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7087e292_a9fd_4420_bb4a_eb2d50355082.slice/crio-96c92813c3c463f20bb4b59d7c2ca49ce9e599f1e78b3e0f5e8d44723cca1e02 WatchSource:0}: Error finding container 96c92813c3c463f20bb4b59d7c2ca49ce9e599f1e78b3e0f5e8d44723cca1e02: Status 404 returned error can't find the container with id 96c92813c3c463f20bb4b59d7c2ca49ce9e599f1e78b3e0f5e8d44723cca1e02 Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.271154 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.387622 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9988868b-f38f-4e95-af61-1834e0e67a8b","Type":"ContainerStarted","Data":"a998b66be8a88486c735874ab97dfbe2f20f83669597b82a6223fde37b4da589"} Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.392458 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7087e292-a9fd-4420-bb4a-eb2d50355082","Type":"ContainerStarted","Data":"96c92813c3c463f20bb4b59d7c2ca49ce9e599f1e78b3e0f5e8d44723cca1e02"} Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.401684 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" event={"ID":"4b728cb5-fb6d-48e7-b170-238f76b6ec51","Type":"ContainerStarted","Data":"2be16b911a0e896e9609cd8da57352116dc02aa05aa515c029c16ec312c61635"} Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.423178 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.425401 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.438823 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.438934 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.439148 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xbg84" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.442298 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.446029 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.453271 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.597011 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-config-data-default\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.597082 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.597137 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.597172 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.597203 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l9wq\" (UniqueName: \"kubernetes.io/projected/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-kube-api-access-9l9wq\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.597231 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-kolla-config\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.597260 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.597295 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.699600 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.699691 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-config-data-default\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.699719 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.699785 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.699820 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.699844 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l9wq\" (UniqueName: \"kubernetes.io/projected/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-kube-api-access-9l9wq\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.699879 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-kolla-config\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.699908 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.701343 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-config-data-generated\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.701662 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-operator-scripts\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.702019 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-kolla-config\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.702630 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.706279 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-config-data-default\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.709362 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.710584 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.720623 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l9wq\" (UniqueName: \"kubernetes.io/projected/6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96-kube-api-access-9l9wq\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.731600 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96\") " pod="openstack/openstack-galera-0" Dec 03 08:58:19 crc kubenswrapper[4573]: I1203 08:58:19.795588 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.503752 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.661359 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.734357 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.734523 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.737967 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-sxb7x" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.738299 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.738562 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.738985 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.837485 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.837561 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.837605 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.837643 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.837676 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.837746 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.837770 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvqpv\" (UniqueName: \"kubernetes.io/projected/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-kube-api-access-jvqpv\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.837804 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.920061 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.921160 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.929664 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-82fwj" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.929949 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.930108 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.938517 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.940350 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.940404 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvqpv\" (UniqueName: \"kubernetes.io/projected/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-kube-api-access-jvqpv\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.940439 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.940502 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.940552 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.940581 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.940624 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.940682 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.942011 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.942669 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.943022 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.943889 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.945122 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.960307 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.966831 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.983526 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvqpv\" (UniqueName: \"kubernetes.io/projected/9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d-kube-api-access-jvqpv\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:20 crc kubenswrapper[4573]: I1203 08:58:20.995242 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.042958 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbnnx\" (UniqueName: \"kubernetes.io/projected/a7d45736-034b-4537-ba45-dd60f3bc3e7a-kube-api-access-xbnnx\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.047760 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7d45736-034b-4537-ba45-dd60f3bc3e7a-kolla-config\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.048010 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7d45736-034b-4537-ba45-dd60f3bc3e7a-config-data\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.048160 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7d45736-034b-4537-ba45-dd60f3bc3e7a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.048441 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d45736-034b-4537-ba45-dd60f3bc3e7a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.079097 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.149672 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbnnx\" (UniqueName: \"kubernetes.io/projected/a7d45736-034b-4537-ba45-dd60f3bc3e7a-kube-api-access-xbnnx\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.149729 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7d45736-034b-4537-ba45-dd60f3bc3e7a-kolla-config\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.149753 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7d45736-034b-4537-ba45-dd60f3bc3e7a-config-data\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.149772 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7d45736-034b-4537-ba45-dd60f3bc3e7a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.149842 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d45736-034b-4537-ba45-dd60f3bc3e7a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.151154 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7d45736-034b-4537-ba45-dd60f3bc3e7a-config-data\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.151206 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a7d45736-034b-4537-ba45-dd60f3bc3e7a-kolla-config\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.157597 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7d45736-034b-4537-ba45-dd60f3bc3e7a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.157714 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7d45736-034b-4537-ba45-dd60f3bc3e7a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.207393 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbnnx\" (UniqueName: \"kubernetes.io/projected/a7d45736-034b-4537-ba45-dd60f3bc3e7a-kube-api-access-xbnnx\") pod \"memcached-0\" (UID: \"a7d45736-034b-4537-ba45-dd60f3bc3e7a\") " pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.260793 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.433988 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96","Type":"ContainerStarted","Data":"a74e1c7507cda3171e28d4e0969e39fa6737335685029f944d5e91c5180c8136"} Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.581208 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 08:58:21 crc kubenswrapper[4573]: I1203 08:58:21.913976 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 08:58:23 crc kubenswrapper[4573]: I1203 08:58:23.447862 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 08:58:23 crc kubenswrapper[4573]: I1203 08:58:23.450806 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 08:58:23 crc kubenswrapper[4573]: I1203 08:58:23.463964 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-867hn" Dec 03 08:58:23 crc kubenswrapper[4573]: I1203 08:58:23.481590 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 08:58:23 crc kubenswrapper[4573]: I1203 08:58:23.514361 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcfpr\" (UniqueName: \"kubernetes.io/projected/811326ff-3ff6-4d68-9c4f-4c356f09d6b2-kube-api-access-qcfpr\") pod \"kube-state-metrics-0\" (UID: \"811326ff-3ff6-4d68-9c4f-4c356f09d6b2\") " pod="openstack/kube-state-metrics-0" Dec 03 08:58:23 crc kubenswrapper[4573]: I1203 08:58:23.616961 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcfpr\" (UniqueName: \"kubernetes.io/projected/811326ff-3ff6-4d68-9c4f-4c356f09d6b2-kube-api-access-qcfpr\") pod \"kube-state-metrics-0\" (UID: \"811326ff-3ff6-4d68-9c4f-4c356f09d6b2\") " pod="openstack/kube-state-metrics-0" Dec 03 08:58:23 crc kubenswrapper[4573]: I1203 08:58:23.644132 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcfpr\" (UniqueName: \"kubernetes.io/projected/811326ff-3ff6-4d68-9c4f-4c356f09d6b2-kube-api-access-qcfpr\") pod \"kube-state-metrics-0\" (UID: \"811326ff-3ff6-4d68-9c4f-4c356f09d6b2\") " pod="openstack/kube-state-metrics-0" Dec 03 08:58:23 crc kubenswrapper[4573]: I1203 08:58:23.792277 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.311642 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.318994 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.322000 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.322619 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.322845 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.322857 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-d5p9t" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.323012 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.328226 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.339816 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.339882 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.339913 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1604082b-b77f-4df2-bea3-ab7803b3e5ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.339957 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1604082b-b77f-4df2-bea3-ab7803b3e5ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.339994 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1604082b-b77f-4df2-bea3-ab7803b3e5ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.340033 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.340090 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8crz\" (UniqueName: \"kubernetes.io/projected/1604082b-b77f-4df2-bea3-ab7803b3e5ec-kube-api-access-c8crz\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.340144 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.442252 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.442655 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.442686 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.442712 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1604082b-b77f-4df2-bea3-ab7803b3e5ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.442753 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1604082b-b77f-4df2-bea3-ab7803b3e5ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.442785 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1604082b-b77f-4df2-bea3-ab7803b3e5ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.442819 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.442851 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8crz\" (UniqueName: \"kubernetes.io/projected/1604082b-b77f-4df2-bea3-ab7803b3e5ec-kube-api-access-c8crz\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.443329 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.443479 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1604082b-b77f-4df2-bea3-ab7803b3e5ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.444108 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1604082b-b77f-4df2-bea3-ab7803b3e5ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.444116 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1604082b-b77f-4df2-bea3-ab7803b3e5ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.449842 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.454889 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.457351 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1604082b-b77f-4df2-bea3-ab7803b3e5ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.461206 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8crz\" (UniqueName: \"kubernetes.io/projected/1604082b-b77f-4df2-bea3-ab7803b3e5ec-kube-api-access-c8crz\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.480400 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1604082b-b77f-4df2-bea3-ab7803b3e5ec\") " pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: W1203 08:58:26.558199 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9671b52a_6c34_4d7d_ab90_fa5b3ef4f89d.slice/crio-9014e878ee977c49b939a3e925706b3861adc56f0e7a75c41130493e7c951e42 WatchSource:0}: Error finding container 9014e878ee977c49b939a3e925706b3861adc56f0e7a75c41130493e7c951e42: Status 404 returned error can't find the container with id 9014e878ee977c49b939a3e925706b3861adc56f0e7a75c41130493e7c951e42 Dec 03 08:58:26 crc kubenswrapper[4573]: W1203 08:58:26.559737 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7d45736_034b_4537_ba45_dd60f3bc3e7a.slice/crio-0e5db1b7969545d08a7099a498fda86792384c4819f4e731b52154e8f1e68717 WatchSource:0}: Error finding container 0e5db1b7969545d08a7099a498fda86792384c4819f4e731b52154e8f1e68717: Status 404 returned error can't find the container with id 0e5db1b7969545d08a7099a498fda86792384c4819f4e731b52154e8f1e68717 Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.649952 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.830991 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bj74b"] Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.835935 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.844480 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.844797 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.849114 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-xk9j8" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.851388 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bj74b"] Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.852861 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-run-ovn\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.858624 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-log-ovn\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.858885 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-run\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.859111 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a7f303-d65a-4546-bbb5-1e223d48d847-ovn-controller-tls-certs\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.859214 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mnbd\" (UniqueName: \"kubernetes.io/projected/33a7f303-d65a-4546-bbb5-1e223d48d847-kube-api-access-7mnbd\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.859622 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a7f303-d65a-4546-bbb5-1e223d48d847-combined-ca-bundle\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.859774 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33a7f303-d65a-4546-bbb5-1e223d48d847-scripts\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.930892 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-jbx79"] Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.943807 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.943885 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.944326 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.953275 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jbx79"] Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.962088 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-run\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.962178 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a7f303-d65a-4546-bbb5-1e223d48d847-ovn-controller-tls-certs\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.962212 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mnbd\" (UniqueName: \"kubernetes.io/projected/33a7f303-d65a-4546-bbb5-1e223d48d847-kube-api-access-7mnbd\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.962279 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvspm\" (UniqueName: \"kubernetes.io/projected/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-kube-api-access-pvspm\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.962323 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a7f303-d65a-4546-bbb5-1e223d48d847-combined-ca-bundle\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.962371 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33a7f303-d65a-4546-bbb5-1e223d48d847-scripts\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.962413 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-scripts\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.964904 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-run\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.965669 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-lib\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.965835 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-log\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.965956 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-etc-ovs\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.966179 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-run\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.966287 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-run-ovn\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.966492 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-log-ovn\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.966571 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-run-ovn\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.966720 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/33a7f303-d65a-4546-bbb5-1e223d48d847-var-log-ovn\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.967555 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33a7f303-d65a-4546-bbb5-1e223d48d847-scripts\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.994782 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33a7f303-d65a-4546-bbb5-1e223d48d847-combined-ca-bundle\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.995071 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mnbd\" (UniqueName: \"kubernetes.io/projected/33a7f303-d65a-4546-bbb5-1e223d48d847-kube-api-access-7mnbd\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:26 crc kubenswrapper[4573]: I1203 08:58:26.998745 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/33a7f303-d65a-4546-bbb5-1e223d48d847-ovn-controller-tls-certs\") pod \"ovn-controller-bj74b\" (UID: \"33a7f303-d65a-4546-bbb5-1e223d48d847\") " pod="openstack/ovn-controller-bj74b" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.069136 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvspm\" (UniqueName: \"kubernetes.io/projected/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-kube-api-access-pvspm\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.069232 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-scripts\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.069274 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-lib\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.069301 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-log\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.069333 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-etc-ovs\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.069372 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-run\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.069609 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-run\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.069835 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-lib\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.070069 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-var-log\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.070250 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-etc-ovs\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.072401 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-scripts\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.098831 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvspm\" (UniqueName: \"kubernetes.io/projected/651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8-kube-api-access-pvspm\") pod \"ovn-controller-ovs-jbx79\" (UID: \"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8\") " pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.179467 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.305303 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.553239 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d","Type":"ContainerStarted","Data":"9014e878ee977c49b939a3e925706b3861adc56f0e7a75c41130493e7c951e42"} Dec 03 08:58:27 crc kubenswrapper[4573]: I1203 08:58:27.567408 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a7d45736-034b-4537-ba45-dd60f3bc3e7a","Type":"ContainerStarted","Data":"0e5db1b7969545d08a7099a498fda86792384c4819f4e731b52154e8f1e68717"} Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.208664 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.212672 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.221625 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.221651 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.221973 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-2wbm9" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.229897 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.251589 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.356861 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5af474c-26c9-4877-989a-911cd77862ab-config\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.356921 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5af474c-26c9-4877-989a-911cd77862ab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.356963 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.356994 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.357122 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5af474c-26c9-4877-989a-911cd77862ab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.357156 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.357186 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q892d\" (UniqueName: \"kubernetes.io/projected/f5af474c-26c9-4877-989a-911cd77862ab-kube-api-access-q892d\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.357224 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459041 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5af474c-26c9-4877-989a-911cd77862ab-config\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459119 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5af474c-26c9-4877-989a-911cd77862ab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459151 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459179 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459241 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5af474c-26c9-4877-989a-911cd77862ab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459269 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459299 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q892d\" (UniqueName: \"kubernetes.io/projected/f5af474c-26c9-4877-989a-911cd77862ab-kube-api-access-q892d\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459336 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.459821 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.468119 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5af474c-26c9-4877-989a-911cd77862ab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.468726 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5af474c-26c9-4877-989a-911cd77862ab-config\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.469117 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5af474c-26c9-4877-989a-911cd77862ab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.480340 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.480356 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.493569 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5af474c-26c9-4877-989a-911cd77862ab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.519967 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.555604 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q892d\" (UniqueName: \"kubernetes.io/projected/f5af474c-26c9-4877-989a-911cd77862ab-kube-api-access-q892d\") pod \"ovsdbserver-sb-0\" (UID: \"f5af474c-26c9-4877-989a-911cd77862ab\") " pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:30 crc kubenswrapper[4573]: I1203 08:58:30.844567 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 08:58:33 crc kubenswrapper[4573]: I1203 08:58:33.541963 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 08:58:41 crc kubenswrapper[4573]: I1203 08:58:41.723723 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"811326ff-3ff6-4d68-9c4f-4c356f09d6b2","Type":"ContainerStarted","Data":"4526b531e75fefcfc7a00490340faba2f23a2806e9df8ac46c1a37937b6c02e4"} Dec 03 08:58:42 crc kubenswrapper[4573]: E1203 08:58:42.928702 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 08:58:42 crc kubenswrapper[4573]: E1203 08:58:42.929385 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mjtl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-b7qxz_openstack(4b728cb5-fb6d-48e7-b170-238f76b6ec51): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:58:42 crc kubenswrapper[4573]: E1203 08:58:42.931310 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" Dec 03 08:58:42 crc kubenswrapper[4573]: E1203 08:58:42.965429 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 08:58:42 crc kubenswrapper[4573]: E1203 08:58:42.965619 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fhdch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-6p7tr_openstack(1010ea0b-292b-4e97-b46d-e3a9794530b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:58:42 crc kubenswrapper[4573]: E1203 08:58:42.966942 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" podUID="1010ea0b-292b-4e97-b46d-e3a9794530b0" Dec 03 08:58:43 crc kubenswrapper[4573]: E1203 08:58:43.051555 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 08:58:43 crc kubenswrapper[4573]: E1203 08:58:43.052198 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nhqw9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-h7ncg_openstack(48ae045b-472f-4957-83ba-5020dde8cde0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:58:43 crc kubenswrapper[4573]: E1203 08:58:43.053739 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" podUID="48ae045b-472f-4957-83ba-5020dde8cde0" Dec 03 08:58:43 crc kubenswrapper[4573]: E1203 08:58:43.760019 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.244260 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.247156 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.422522 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48ae045b-472f-4957-83ba-5020dde8cde0-config\") pod \"48ae045b-472f-4957-83ba-5020dde8cde0\" (UID: \"48ae045b-472f-4957-83ba-5020dde8cde0\") " Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.423037 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/48ae045b-472f-4957-83ba-5020dde8cde0-kube-api-access-nhqw9\") pod \"48ae045b-472f-4957-83ba-5020dde8cde0\" (UID: \"48ae045b-472f-4957-83ba-5020dde8cde0\") " Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.423250 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhdch\" (UniqueName: \"kubernetes.io/projected/1010ea0b-292b-4e97-b46d-e3a9794530b0-kube-api-access-fhdch\") pod \"1010ea0b-292b-4e97-b46d-e3a9794530b0\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.423248 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48ae045b-472f-4957-83ba-5020dde8cde0-config" (OuterVolumeSpecName: "config") pod "48ae045b-472f-4957-83ba-5020dde8cde0" (UID: "48ae045b-472f-4957-83ba-5020dde8cde0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.423282 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-config\") pod \"1010ea0b-292b-4e97-b46d-e3a9794530b0\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.423303 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-dns-svc\") pod \"1010ea0b-292b-4e97-b46d-e3a9794530b0\" (UID: \"1010ea0b-292b-4e97-b46d-e3a9794530b0\") " Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.424001 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48ae045b-472f-4957-83ba-5020dde8cde0-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.425263 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-config" (OuterVolumeSpecName: "config") pod "1010ea0b-292b-4e97-b46d-e3a9794530b0" (UID: "1010ea0b-292b-4e97-b46d-e3a9794530b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.427811 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1010ea0b-292b-4e97-b46d-e3a9794530b0" (UID: "1010ea0b-292b-4e97-b46d-e3a9794530b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.428426 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1010ea0b-292b-4e97-b46d-e3a9794530b0-kube-api-access-fhdch" (OuterVolumeSpecName: "kube-api-access-fhdch") pod "1010ea0b-292b-4e97-b46d-e3a9794530b0" (UID: "1010ea0b-292b-4e97-b46d-e3a9794530b0"). InnerVolumeSpecName "kube-api-access-fhdch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.430944 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48ae045b-472f-4957-83ba-5020dde8cde0-kube-api-access-nhqw9" (OuterVolumeSpecName: "kube-api-access-nhqw9") pod "48ae045b-472f-4957-83ba-5020dde8cde0" (UID: "48ae045b-472f-4957-83ba-5020dde8cde0"). InnerVolumeSpecName "kube-api-access-nhqw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.527351 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhdch\" (UniqueName: \"kubernetes.io/projected/1010ea0b-292b-4e97-b46d-e3a9794530b0-kube-api-access-fhdch\") on node \"crc\" DevicePath \"\"" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.527399 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.527417 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1010ea0b-292b-4e97-b46d-e3a9794530b0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.527432 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhqw9\" (UniqueName: \"kubernetes.io/projected/48ae045b-472f-4957-83ba-5020dde8cde0-kube-api-access-nhqw9\") on node \"crc\" DevicePath \"\"" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.805089 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" event={"ID":"48ae045b-472f-4957-83ba-5020dde8cde0","Type":"ContainerDied","Data":"12abb32a81bb0958a339b63e6e53a43643dce4287cc082c53ad575a992774817"} Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.805178 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h7ncg" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.809813 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" event={"ID":"1010ea0b-292b-4e97-b46d-e3a9794530b0","Type":"ContainerDied","Data":"419f5ed0dc3ef971d4e612fb98a434a5ecb3e42c17de13811379628d2696acd7"} Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.809976 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-6p7tr" Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.883531 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.949829 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6p7tr"] Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.967800 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bj74b"] Dec 03 08:58:46 crc kubenswrapper[4573]: I1203 08:58:46.992989 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-6p7tr"] Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.019716 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7ncg"] Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.028902 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7ncg"] Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.615293 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.878323 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b" event={"ID":"33a7f303-d65a-4546-bbb5-1e223d48d847","Type":"ContainerStarted","Data":"3b92be41a7b9f2a2d7eff5ef4f4c9332cbae0ca4b24b0e7b226d9fc7bc13c2da"} Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.880734 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7087e292-a9fd-4420-bb4a-eb2d50355082","Type":"ContainerStarted","Data":"5f27bfda26561875fed220674d65c66eac1bdc140eca3aa513d193416a640c55"} Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.889683 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-jbx79"] Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.890155 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f5af474c-26c9-4877-989a-911cd77862ab","Type":"ContainerStarted","Data":"3d7192cda2ab338cf9be378ae32322d3b4bfce01acfd80e9a8c00cc9f7c313b2"} Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.906473 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1604082b-b77f-4df2-bea3-ab7803b3e5ec","Type":"ContainerStarted","Data":"55fa793e1f26f635cddeb28f5f54144811c39712949f55acd748552122430aff"} Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.930677 4573 generic.go:334] "Generic (PLEG): container finished" podID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" containerID="db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0" exitCode=0 Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.930757 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" event={"ID":"12f89059-a1f2-45fa-be11-e4bdc6b6f936","Type":"ContainerDied","Data":"db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0"} Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.935759 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 08:58:47 crc kubenswrapper[4573]: I1203 08:58:47.980028 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=8.40858149 podStartE2EDuration="27.980000614s" podCreationTimestamp="2025-12-03 08:58:20 +0000 UTC" firstStartedPulling="2025-12-03 08:58:26.579667118 +0000 UTC m=+1227.148046377" lastFinishedPulling="2025-12-03 08:58:46.151086242 +0000 UTC m=+1246.719465501" observedRunningTime="2025-12-03 08:58:47.978332118 +0000 UTC m=+1248.546711377" watchObservedRunningTime="2025-12-03 08:58:47.980000614 +0000 UTC m=+1248.548379873" Dec 03 08:58:48 crc kubenswrapper[4573]: I1203 08:58:48.044095 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1010ea0b-292b-4e97-b46d-e3a9794530b0" path="/var/lib/kubelet/pods/1010ea0b-292b-4e97-b46d-e3a9794530b0/volumes" Dec 03 08:58:48 crc kubenswrapper[4573]: I1203 08:58:48.044571 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48ae045b-472f-4957-83ba-5020dde8cde0" path="/var/lib/kubelet/pods/48ae045b-472f-4957-83ba-5020dde8cde0/volumes" Dec 03 08:58:48 crc kubenswrapper[4573]: I1203 08:58:48.956237 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jbx79" event={"ID":"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8","Type":"ContainerStarted","Data":"50303e4a58a4accc55c3e84df4da1039641800f2ee85be2c382d6b94a80b694e"} Dec 03 08:58:48 crc kubenswrapper[4573]: I1203 08:58:48.963456 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9988868b-f38f-4e95-af61-1834e0e67a8b","Type":"ContainerStarted","Data":"b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca"} Dec 03 08:58:48 crc kubenswrapper[4573]: I1203 08:58:48.976713 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a7d45736-034b-4537-ba45-dd60f3bc3e7a","Type":"ContainerStarted","Data":"3caeee6938e5248e7a80e614db510365b74e2c36ce92e9d340955311af21cfc8"} Dec 03 08:58:48 crc kubenswrapper[4573]: I1203 08:58:48.984549 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" event={"ID":"12f89059-a1f2-45fa-be11-e4bdc6b6f936","Type":"ContainerStarted","Data":"5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5"} Dec 03 08:58:48 crc kubenswrapper[4573]: I1203 08:58:48.984866 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:48 crc kubenswrapper[4573]: I1203 08:58:48.990945 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d","Type":"ContainerStarted","Data":"edc81e0941051775221684b2c10b515a286055992be385a61520b47ce95db3b7"} Dec 03 08:58:49 crc kubenswrapper[4573]: I1203 08:58:48.999431 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96","Type":"ContainerStarted","Data":"81ea4bb68d1f8f52b2dfb14fe119d696664b6d71df7118687a8c9e4a692950d7"} Dec 03 08:58:49 crc kubenswrapper[4573]: I1203 08:58:49.014904 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" podStartSLOduration=5.21859109 podStartE2EDuration="33.014869738s" podCreationTimestamp="2025-12-03 08:58:16 +0000 UTC" firstStartedPulling="2025-12-03 08:58:18.325514132 +0000 UTC m=+1218.893893381" lastFinishedPulling="2025-12-03 08:58:46.12179277 +0000 UTC m=+1246.690172029" observedRunningTime="2025-12-03 08:58:49.006798787 +0000 UTC m=+1249.575178046" watchObservedRunningTime="2025-12-03 08:58:49.014869738 +0000 UTC m=+1249.583248997" Dec 03 08:58:51 crc kubenswrapper[4573]: I1203 08:58:51.025852 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"811326ff-3ff6-4d68-9c4f-4c356f09d6b2","Type":"ContainerStarted","Data":"64ce8654f051565f7b36c0dede7b298fbebd6eb40cf6baef743ad2b103898f4a"} Dec 03 08:58:51 crc kubenswrapper[4573]: I1203 08:58:51.026536 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 08:58:51 crc kubenswrapper[4573]: I1203 08:58:51.048961 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.433674438 podStartE2EDuration="28.048904341s" podCreationTimestamp="2025-12-03 08:58:23 +0000 UTC" firstStartedPulling="2025-12-03 08:58:41.447922716 +0000 UTC m=+1242.016301975" lastFinishedPulling="2025-12-03 08:58:50.063152619 +0000 UTC m=+1250.631531878" observedRunningTime="2025-12-03 08:58:51.0474008 +0000 UTC m=+1251.615780069" watchObservedRunningTime="2025-12-03 08:58:51.048904341 +0000 UTC m=+1251.617283620" Dec 03 08:58:52 crc kubenswrapper[4573]: I1203 08:58:52.042819 4573 generic.go:334] "Generic (PLEG): container finished" podID="9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d" containerID="edc81e0941051775221684b2c10b515a286055992be385a61520b47ce95db3b7" exitCode=0 Dec 03 08:58:52 crc kubenswrapper[4573]: I1203 08:58:52.043085 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d","Type":"ContainerDied","Data":"edc81e0941051775221684b2c10b515a286055992be385a61520b47ce95db3b7"} Dec 03 08:58:53 crc kubenswrapper[4573]: I1203 08:58:53.066615 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jbx79" event={"ID":"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8","Type":"ContainerStarted","Data":"79b9f82b94cb8d2697129af2d60633977290a93d0cc07d0f0a71db6abfdb07dd"} Dec 03 08:58:53 crc kubenswrapper[4573]: I1203 08:58:53.078391 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f5af474c-26c9-4877-989a-911cd77862ab","Type":"ContainerStarted","Data":"14cbd2d264594144d782699b15aff075742c3d24f86b8ca98116c2f4900b5eb3"} Dec 03 08:58:53 crc kubenswrapper[4573]: I1203 08:58:53.081739 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d","Type":"ContainerStarted","Data":"59660e7ebe60574e094c0ddd6a227fb975c3d96871f409fdcfe467c0073c4e15"} Dec 03 08:58:53 crc kubenswrapper[4573]: I1203 08:58:53.088563 4573 generic.go:334] "Generic (PLEG): container finished" podID="6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96" containerID="81ea4bb68d1f8f52b2dfb14fe119d696664b6d71df7118687a8c9e4a692950d7" exitCode=0 Dec 03 08:58:53 crc kubenswrapper[4573]: I1203 08:58:53.088645 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96","Type":"ContainerDied","Data":"81ea4bb68d1f8f52b2dfb14fe119d696664b6d71df7118687a8c9e4a692950d7"} Dec 03 08:58:53 crc kubenswrapper[4573]: I1203 08:58:53.123806 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=14.390285754 podStartE2EDuration="34.12376336s" podCreationTimestamp="2025-12-03 08:58:19 +0000 UTC" firstStartedPulling="2025-12-03 08:58:26.579665938 +0000 UTC m=+1227.148045197" lastFinishedPulling="2025-12-03 08:58:46.313143534 +0000 UTC m=+1246.881522803" observedRunningTime="2025-12-03 08:58:53.121493668 +0000 UTC m=+1253.689872947" watchObservedRunningTime="2025-12-03 08:58:53.12376336 +0000 UTC m=+1253.692142619" Dec 03 08:58:54 crc kubenswrapper[4573]: I1203 08:58:54.122528 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96","Type":"ContainerStarted","Data":"fb209b85c4b262273fcfb524b2187c5675aeb21e8977615b4d9cede82d5a3edb"} Dec 03 08:58:54 crc kubenswrapper[4573]: I1203 08:58:54.125986 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1604082b-b77f-4df2-bea3-ab7803b3e5ec","Type":"ContainerStarted","Data":"0822d20e1f295d4c30312add94057eca0c6ae1130ade6df0b91a2d0b5cb547f1"} Dec 03 08:58:54 crc kubenswrapper[4573]: I1203 08:58:54.131180 4573 generic.go:334] "Generic (PLEG): container finished" podID="651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8" containerID="79b9f82b94cb8d2697129af2d60633977290a93d0cc07d0f0a71db6abfdb07dd" exitCode=0 Dec 03 08:58:54 crc kubenswrapper[4573]: I1203 08:58:54.131281 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jbx79" event={"ID":"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8","Type":"ContainerDied","Data":"79b9f82b94cb8d2697129af2d60633977290a93d0cc07d0f0a71db6abfdb07dd"} Dec 03 08:58:54 crc kubenswrapper[4573]: I1203 08:58:54.144332 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b" event={"ID":"33a7f303-d65a-4546-bbb5-1e223d48d847","Type":"ContainerStarted","Data":"508c40a9e417824a7c66f0c8929b08871543478821c41cf69983f5b741eff35f"} Dec 03 08:58:54 crc kubenswrapper[4573]: I1203 08:58:54.144575 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-bj74b" Dec 03 08:58:54 crc kubenswrapper[4573]: I1203 08:58:54.162740 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=10.381868161 podStartE2EDuration="36.162713807s" podCreationTimestamp="2025-12-03 08:58:18 +0000 UTC" firstStartedPulling="2025-12-03 08:58:20.532287878 +0000 UTC m=+1221.100667137" lastFinishedPulling="2025-12-03 08:58:46.313133524 +0000 UTC m=+1246.881512783" observedRunningTime="2025-12-03 08:58:54.155876259 +0000 UTC m=+1254.724255548" watchObservedRunningTime="2025-12-03 08:58:54.162713807 +0000 UTC m=+1254.731093066" Dec 03 08:58:54 crc kubenswrapper[4573]: I1203 08:58:54.183702 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-bj74b" podStartSLOduration=23.249127796 podStartE2EDuration="28.18367333s" podCreationTimestamp="2025-12-03 08:58:26 +0000 UTC" firstStartedPulling="2025-12-03 08:58:47.762093104 +0000 UTC m=+1248.330472363" lastFinishedPulling="2025-12-03 08:58:52.696638638 +0000 UTC m=+1253.265017897" observedRunningTime="2025-12-03 08:58:54.181590972 +0000 UTC m=+1254.749970241" watchObservedRunningTime="2025-12-03 08:58:54.18367333 +0000 UTC m=+1254.752052589" Dec 03 08:58:55 crc kubenswrapper[4573]: I1203 08:58:55.169856 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jbx79" event={"ID":"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8","Type":"ContainerStarted","Data":"ca26b5c17ca43bc3cb620b0a1d2406778e5da6ba0f943dd2706fbc8b7cbb8ecd"} Dec 03 08:58:55 crc kubenswrapper[4573]: I1203 08:58:55.170190 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-jbx79" event={"ID":"651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8","Type":"ContainerStarted","Data":"eafd15dd4a1c037f580bc309c252eaf0675c1407731ed07923fbebd4d352442d"} Dec 03 08:58:55 crc kubenswrapper[4573]: I1203 08:58:55.170225 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:55 crc kubenswrapper[4573]: I1203 08:58:55.170245 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:58:55 crc kubenswrapper[4573]: I1203 08:58:55.251802 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-jbx79" podStartSLOduration=24.754763075 podStartE2EDuration="29.251772753s" podCreationTimestamp="2025-12-03 08:58:26 +0000 UTC" firstStartedPulling="2025-12-03 08:58:48.160943513 +0000 UTC m=+1248.729322772" lastFinishedPulling="2025-12-03 08:58:52.657953191 +0000 UTC m=+1253.226332450" observedRunningTime="2025-12-03 08:58:55.238751727 +0000 UTC m=+1255.807130986" watchObservedRunningTime="2025-12-03 08:58:55.251772753 +0000 UTC m=+1255.820152012" Dec 03 08:58:56 crc kubenswrapper[4573]: I1203 08:58:56.262924 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 08:58:56 crc kubenswrapper[4573]: I1203 08:58:56.943369 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 08:58:56 crc kubenswrapper[4573]: I1203 08:58:56.943818 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 08:58:56 crc kubenswrapper[4573]: I1203 08:58:56.943910 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 08:58:56 crc kubenswrapper[4573]: I1203 08:58:56.944897 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93ffc1f0cd99ae07290bc9e27b4a1d6c063abbb15b748bb21e1aa18b0c38a447"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 08:58:56 crc kubenswrapper[4573]: I1203 08:58:56.944982 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://93ffc1f0cd99ae07290bc9e27b4a1d6c063abbb15b748bb21e1aa18b0c38a447" gracePeriod=600 Dec 03 08:58:57 crc kubenswrapper[4573]: I1203 08:58:57.271597 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:58:58 crc kubenswrapper[4573]: I1203 08:58:58.196019 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f5af474c-26c9-4877-989a-911cd77862ab","Type":"ContainerStarted","Data":"c204013f027f1549184b23e569fdccacec198d3a5ea1a78d3cf90eac4dc2d11f"} Dec 03 08:58:58 crc kubenswrapper[4573]: I1203 08:58:58.199315 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1604082b-b77f-4df2-bea3-ab7803b3e5ec","Type":"ContainerStarted","Data":"6a72f2ee55cece1cedd38ea4ba6e1118847d8218249d453dc8ec6ea5afbe0dde"} Dec 03 08:58:58 crc kubenswrapper[4573]: I1203 08:58:58.202033 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="93ffc1f0cd99ae07290bc9e27b4a1d6c063abbb15b748bb21e1aa18b0c38a447" exitCode=0 Dec 03 08:58:58 crc kubenswrapper[4573]: I1203 08:58:58.202106 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"93ffc1f0cd99ae07290bc9e27b4a1d6c063abbb15b748bb21e1aa18b0c38a447"} Dec 03 08:58:58 crc kubenswrapper[4573]: I1203 08:58:58.202174 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"2b3b734897ed0efe44e3ec39a1d49da70613f4f9958b9ef5aee9f0572f629605"} Dec 03 08:58:58 crc kubenswrapper[4573]: I1203 08:58:58.202226 4573 scope.go:117] "RemoveContainer" containerID="68d9133084811f469a486ce4e9540b723fb96d8cd528443eac595b0a1c6329ed" Dec 03 08:58:58 crc kubenswrapper[4573]: I1203 08:58:58.230648 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=19.452424754 podStartE2EDuration="29.230625627s" podCreationTimestamp="2025-12-03 08:58:29 +0000 UTC" firstStartedPulling="2025-12-03 08:58:47.659456826 +0000 UTC m=+1248.227836085" lastFinishedPulling="2025-12-03 08:58:57.437657699 +0000 UTC m=+1258.006036958" observedRunningTime="2025-12-03 08:58:58.228644043 +0000 UTC m=+1258.797023322" watchObservedRunningTime="2025-12-03 08:58:58.230625627 +0000 UTC m=+1258.799004906" Dec 03 08:58:58 crc kubenswrapper[4573]: I1203 08:58:58.264382 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=23.623537665 podStartE2EDuration="33.26436429s" podCreationTimestamp="2025-12-03 08:58:25 +0000 UTC" firstStartedPulling="2025-12-03 08:58:47.788954428 +0000 UTC m=+1248.357333687" lastFinishedPulling="2025-12-03 08:58:57.429781053 +0000 UTC m=+1257.998160312" observedRunningTime="2025-12-03 08:58:58.256602808 +0000 UTC m=+1258.824982067" watchObservedRunningTime="2025-12-03 08:58:58.26436429 +0000 UTC m=+1258.832743549" Dec 03 08:58:59 crc kubenswrapper[4573]: I1203 08:58:59.650745 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:59 crc kubenswrapper[4573]: I1203 08:58:59.695094 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 08:58:59 crc kubenswrapper[4573]: E1203 08:58:59.779976 4573 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.251:39208->38.102.83.251:37631: read tcp 38.102.83.251:39208->38.102.83.251:37631: read: connection reset by peer Dec 03 08:58:59 crc kubenswrapper[4573]: I1203 08:58:59.796749 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 08:58:59 crc kubenswrapper[4573]: I1203 08:58:59.796830 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 08:58:59 crc kubenswrapper[4573]: I1203 08:58:59.909431 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.375952 4573 generic.go:334] "Generic (PLEG): container finished" podID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" containerID="ed5f5a3c9f995f9aebe30e238d48f500aa8378deb069dd20fc1d4a112f219619" exitCode=0 Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.376038 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" event={"ID":"4b728cb5-fb6d-48e7-b170-238f76b6ec51","Type":"ContainerDied","Data":"ed5f5a3c9f995f9aebe30e238d48f500aa8378deb069dd20fc1d4a112f219619"} Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.376498 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.477419 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.576349 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.826801 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b7qxz"] Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.845336 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.845446 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.883972 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-dgch8"] Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.885802 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.890631 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.920158 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-dgch8"] Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.923061 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7dt9\" (UniqueName: \"kubernetes.io/projected/0cf7deca-25aa-4516-869a-ad2a20419d3a-kube-api-access-m7dt9\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.923153 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.923292 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.923333 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-config\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:00 crc kubenswrapper[4573]: I1203 08:59:00.941201 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.040280 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7dt9\" (UniqueName: \"kubernetes.io/projected/0cf7deca-25aa-4516-869a-ad2a20419d3a-kube-api-access-m7dt9\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.040337 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.040411 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.040438 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-config\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.041860 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.043106 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-config\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.043208 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.056282 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-74eb-account-create-update-hhzh4"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.057713 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.062939 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.081031 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.082097 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.124370 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7dt9\" (UniqueName: \"kubernetes.io/projected/0cf7deca-25aa-4516-869a-ad2a20419d3a-kube-api-access-m7dt9\") pod \"dnsmasq-dns-5bf47b49b7-dgch8\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.205056 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.243649 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-6p2pk"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.254763 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d9684c-2d7b-49df-a5c6-a56656a84874-operator-scripts\") pod \"keystone-74eb-account-create-update-hhzh4\" (UID: \"81d9684c-2d7b-49df-a5c6-a56656a84874\") " pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.254869 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44vhh\" (UniqueName: \"kubernetes.io/projected/81d9684c-2d7b-49df-a5c6-a56656a84874-kube-api-access-44vhh\") pod \"keystone-74eb-account-create-update-hhzh4\" (UID: \"81d9684c-2d7b-49df-a5c6-a56656a84874\") " pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.257146 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.333309 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-6p2pk"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.354109 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-rrcpn"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.355742 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.360119 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.370060 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d9684c-2d7b-49df-a5c6-a56656a84874-operator-scripts\") pod \"keystone-74eb-account-create-update-hhzh4\" (UID: \"81d9684c-2d7b-49df-a5c6-a56656a84874\") " pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.375312 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rrcpn"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.382602 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cblvq\" (UniqueName: \"kubernetes.io/projected/f0d549a7-b237-46de-8dde-d7aba95cfe65-kube-api-access-cblvq\") pod \"keystone-db-create-6p2pk\" (UID: \"f0d549a7-b237-46de-8dde-d7aba95cfe65\") " pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.382680 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44vhh\" (UniqueName: \"kubernetes.io/projected/81d9684c-2d7b-49df-a5c6-a56656a84874-kube-api-access-44vhh\") pod \"keystone-74eb-account-create-update-hhzh4\" (UID: \"81d9684c-2d7b-49df-a5c6-a56656a84874\") " pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.382895 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0d549a7-b237-46de-8dde-d7aba95cfe65-operator-scripts\") pod \"keystone-db-create-6p2pk\" (UID: \"f0d549a7-b237-46de-8dde-d7aba95cfe65\") " pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.382901 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d9684c-2d7b-49df-a5c6-a56656a84874-operator-scripts\") pod \"keystone-74eb-account-create-update-hhzh4\" (UID: \"81d9684c-2d7b-49df-a5c6-a56656a84874\") " pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.394495 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-74eb-account-create-update-hhzh4"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.408465 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44vhh\" (UniqueName: \"kubernetes.io/projected/81d9684c-2d7b-49df-a5c6-a56656a84874-kube-api-access-44vhh\") pod \"keystone-74eb-account-create-update-hhzh4\" (UID: \"81d9684c-2d7b-49df-a5c6-a56656a84874\") " pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.414663 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-nfl5h"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.417807 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.429667 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nfl5h"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.451746 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" containerName="dnsmasq-dns" containerID="cri-o://e6f06726d144ccb117ad2ea1372333c2a0959919e71f9719caa0435ec7809e20" gracePeriod=10 Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.452145 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" event={"ID":"4b728cb5-fb6d-48e7-b170-238f76b6ec51","Type":"ContainerStarted","Data":"e6f06726d144ccb117ad2ea1372333c2a0959919e71f9719caa0435ec7809e20"} Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.453334 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.462182 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-e92a-account-create-update-6bjzf"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.463556 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.474014 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.485281 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-config\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.485338 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vphms\" (UniqueName: \"kubernetes.io/projected/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-kube-api-access-vphms\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.485395 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.485420 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-combined-ca-bundle\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.485489 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-ovn-rundir\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.485540 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cblvq\" (UniqueName: \"kubernetes.io/projected/f0d549a7-b237-46de-8dde-d7aba95cfe65-kube-api-access-cblvq\") pod \"keystone-db-create-6p2pk\" (UID: \"f0d549a7-b237-46de-8dde-d7aba95cfe65\") " pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.485605 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-ovs-rundir\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.485659 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0d549a7-b237-46de-8dde-d7aba95cfe65-operator-scripts\") pod \"keystone-db-create-6p2pk\" (UID: \"f0d549a7-b237-46de-8dde-d7aba95cfe65\") " pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.486640 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0d549a7-b237-46de-8dde-d7aba95cfe65-operator-scripts\") pod \"keystone-db-create-6p2pk\" (UID: \"f0d549a7-b237-46de-8dde-d7aba95cfe65\") " pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.518060 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cblvq\" (UniqueName: \"kubernetes.io/projected/f0d549a7-b237-46de-8dde-d7aba95cfe65-kube-api-access-cblvq\") pod \"keystone-db-create-6p2pk\" (UID: \"f0d549a7-b237-46de-8dde-d7aba95cfe65\") " pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.575138 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e92a-account-create-update-6bjzf"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.599172 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpnbq\" (UniqueName: \"kubernetes.io/projected/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-kube-api-access-lpnbq\") pod \"placement-db-create-nfl5h\" (UID: \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\") " pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.599355 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-ovn-rundir\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.599585 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-operator-scripts\") pod \"placement-db-create-nfl5h\" (UID: \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\") " pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.599696 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-ovs-rundir\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.599784 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62wtm\" (UniqueName: \"kubernetes.io/projected/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-kube-api-access-62wtm\") pod \"placement-e92a-account-create-update-6bjzf\" (UID: \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\") " pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.599851 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-ovn-rundir\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.599937 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-operator-scripts\") pod \"placement-e92a-account-create-update-6bjzf\" (UID: \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\") " pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.601098 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-ovs-rundir\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.602304 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-config\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.602380 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vphms\" (UniqueName: \"kubernetes.io/projected/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-kube-api-access-vphms\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.602450 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.602474 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-combined-ca-bundle\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.603227 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-config\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.607554 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.614392 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.677009 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vphms\" (UniqueName: \"kubernetes.io/projected/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-kube-api-access-vphms\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.677599 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.683187 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.685791 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a19cac1-e9f7-4f7d-b356-3263c1b8c446-combined-ca-bundle\") pod \"ovn-controller-metrics-rrcpn\" (UID: \"6a19cac1-e9f7-4f7d-b356-3263c1b8c446\") " pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.704508 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpnbq\" (UniqueName: \"kubernetes.io/projected/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-kube-api-access-lpnbq\") pod \"placement-db-create-nfl5h\" (UID: \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\") " pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.711875 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-operator-scripts\") pod \"placement-db-create-nfl5h\" (UID: \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\") " pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.712016 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62wtm\" (UniqueName: \"kubernetes.io/projected/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-kube-api-access-62wtm\") pod \"placement-e92a-account-create-update-6bjzf\" (UID: \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\") " pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.712154 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-operator-scripts\") pod \"placement-e92a-account-create-update-6bjzf\" (UID: \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\") " pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.713417 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-operator-scripts\") pod \"placement-e92a-account-create-update-6bjzf\" (UID: \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\") " pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.715171 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-operator-scripts\") pod \"placement-db-create-nfl5h\" (UID: \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\") " pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.729144 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-rrcpn" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.736335 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-dgch8"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.754889 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62wtm\" (UniqueName: \"kubernetes.io/projected/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-kube-api-access-62wtm\") pod \"placement-e92a-account-create-update-6bjzf\" (UID: \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\") " pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.766777 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" podStartSLOduration=-9223371992.113064 podStartE2EDuration="44.741712318s" podCreationTimestamp="2025-12-03 08:58:17 +0000 UTC" firstStartedPulling="2025-12-03 08:58:18.43741451 +0000 UTC m=+1219.005793769" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:01.507632576 +0000 UTC m=+1262.076011835" watchObservedRunningTime="2025-12-03 08:59:01.741712318 +0000 UTC m=+1262.310091577" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.799043 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.800939 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpnbq\" (UniqueName: \"kubernetes.io/projected/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-kube-api-access-lpnbq\") pod \"placement-db-create-nfl5h\" (UID: \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\") " pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.877057 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-pgr62"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.881311 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.891741 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-pgr62"] Dec 03 08:59:01 crc kubenswrapper[4573]: I1203 08:59:01.896579 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.019280 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-dns-svc\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.019319 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.019381 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx4hf\" (UniqueName: \"kubernetes.io/projected/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-kube-api-access-jx4hf\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.019414 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.019479 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-config\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.077113 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.121449 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-config\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.121551 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-dns-svc\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.121569 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.121610 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx4hf\" (UniqueName: \"kubernetes.io/projected/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-kube-api-access-jx4hf\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.121652 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.122589 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.123231 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-dns-svc\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.123800 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.124006 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-config\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.170799 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx4hf\" (UniqueName: \"kubernetes.io/projected/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-kube-api-access-jx4hf\") pod \"dnsmasq-dns-8554648995-pgr62\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.213911 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.282933 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.284516 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.289166 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.289428 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.289586 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-jqvvm" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.289818 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.338056 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.338136 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbk52\" (UniqueName: \"kubernetes.io/projected/1999a198-5df7-40fc-ba0f-5b650b3027a6-kube-api-access-lbk52\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.338204 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1999a198-5df7-40fc-ba0f-5b650b3027a6-scripts\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.338265 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.338330 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1999a198-5df7-40fc-ba0f-5b650b3027a6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.338353 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.338376 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1999a198-5df7-40fc-ba0f-5b650b3027a6-config\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.357942 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.460747 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1999a198-5df7-40fc-ba0f-5b650b3027a6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.461176 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.461324 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1999a198-5df7-40fc-ba0f-5b650b3027a6-config\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.461489 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.461615 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbk52\" (UniqueName: \"kubernetes.io/projected/1999a198-5df7-40fc-ba0f-5b650b3027a6-kube-api-access-lbk52\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.461771 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1999a198-5df7-40fc-ba0f-5b650b3027a6-scripts\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.461937 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.470434 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1999a198-5df7-40fc-ba0f-5b650b3027a6-config\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.489040 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.489299 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1999a198-5df7-40fc-ba0f-5b650b3027a6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.490848 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1999a198-5df7-40fc-ba0f-5b650b3027a6-scripts\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.569244 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.575226 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999a198-5df7-40fc-ba0f-5b650b3027a6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.590450 4573 generic.go:334] "Generic (PLEG): container finished" podID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" containerID="e6f06726d144ccb117ad2ea1372333c2a0959919e71f9719caa0435ec7809e20" exitCode=0 Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.592094 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" event={"ID":"4b728cb5-fb6d-48e7-b170-238f76b6ec51","Type":"ContainerDied","Data":"e6f06726d144ccb117ad2ea1372333c2a0959919e71f9719caa0435ec7809e20"} Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.592139 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" event={"ID":"4b728cb5-fb6d-48e7-b170-238f76b6ec51","Type":"ContainerDied","Data":"2be16b911a0e896e9609cd8da57352116dc02aa05aa515c029c16ec312c61635"} Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.592177 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2be16b911a0e896e9609cd8da57352116dc02aa05aa515c029c16ec312c61635" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.624639 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbk52\" (UniqueName: \"kubernetes.io/projected/1999a198-5df7-40fc-ba0f-5b650b3027a6-kube-api-access-lbk52\") pod \"ovn-northd-0\" (UID: \"1999a198-5df7-40fc-ba0f-5b650b3027a6\") " pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.630751 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-dgch8"] Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.689707 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.695640 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.797095 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjtl8\" (UniqueName: \"kubernetes.io/projected/4b728cb5-fb6d-48e7-b170-238f76b6ec51-kube-api-access-mjtl8\") pod \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.797779 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-dns-svc\") pod \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.797900 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-config\") pod \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\" (UID: \"4b728cb5-fb6d-48e7-b170-238f76b6ec51\") " Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.810602 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b728cb5-fb6d-48e7-b170-238f76b6ec51-kube-api-access-mjtl8" (OuterVolumeSpecName: "kube-api-access-mjtl8") pod "4b728cb5-fb6d-48e7-b170-238f76b6ec51" (UID: "4b728cb5-fb6d-48e7-b170-238f76b6ec51"). InnerVolumeSpecName "kube-api-access-mjtl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.885408 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-config" (OuterVolumeSpecName: "config") pod "4b728cb5-fb6d-48e7-b170-238f76b6ec51" (UID: "4b728cb5-fb6d-48e7-b170-238f76b6ec51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.899107 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4b728cb5-fb6d-48e7-b170-238f76b6ec51" (UID: "4b728cb5-fb6d-48e7-b170-238f76b6ec51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.918286 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.918551 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b728cb5-fb6d-48e7-b170-238f76b6ec51-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.921291 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjtl8\" (UniqueName: \"kubernetes.io/projected/4b728cb5-fb6d-48e7-b170-238f76b6ec51-kube-api-access-mjtl8\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:02 crc kubenswrapper[4573]: I1203 08:59:02.930201 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-6p2pk"] Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.029614 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-74eb-account-create-update-hhzh4"] Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.270803 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e92a-account-create-update-6bjzf"] Dec 03 08:59:03 crc kubenswrapper[4573]: W1203 08:59:03.279773 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode78a67e6_fbd8_4c2b_bc2d_148de6b27c20.slice/crio-c0807268420a67fe4f110c9f1bf5fabeed9719cc45407b2a5897a39fe8c4486a WatchSource:0}: Error finding container c0807268420a67fe4f110c9f1bf5fabeed9719cc45407b2a5897a39fe8c4486a: Status 404 returned error can't find the container with id c0807268420a67fe4f110c9f1bf5fabeed9719cc45407b2a5897a39fe8c4486a Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.296514 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-rrcpn"] Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.328910 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nfl5h"] Dec 03 08:59:03 crc kubenswrapper[4573]: W1203 08:59:03.343010 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd12c3b6_45e8_44dd_b28f_b7c828cb5690.slice/crio-8029c068fbce6aa6701f1dc03f3f350fab15e5a0a82d97c7126282f51c95b531 WatchSource:0}: Error finding container 8029c068fbce6aa6701f1dc03f3f350fab15e5a0a82d97c7126282f51c95b531: Status 404 returned error can't find the container with id 8029c068fbce6aa6701f1dc03f3f350fab15e5a0a82d97c7126282f51c95b531 Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.515010 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-pgr62"] Dec 03 08:59:03 crc kubenswrapper[4573]: W1203 08:59:03.551432 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5fa77a1_fea1_4f30_8076_ad7e4bab23ee.slice/crio-3985626d4c0073f71f868af55f4ae83665d63b16727cd39283a5ab4a453a294d WatchSource:0}: Error finding container 3985626d4c0073f71f868af55f4ae83665d63b16727cd39283a5ab4a453a294d: Status 404 returned error can't find the container with id 3985626d4c0073f71f868af55f4ae83665d63b16727cd39283a5ab4a453a294d Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.627211 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nfl5h" event={"ID":"dd12c3b6-45e8-44dd-b28f-b7c828cb5690","Type":"ContainerStarted","Data":"8029c068fbce6aa6701f1dc03f3f350fab15e5a0a82d97c7126282f51c95b531"} Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.647764 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rrcpn" event={"ID":"6a19cac1-e9f7-4f7d-b356-3263c1b8c446","Type":"ContainerStarted","Data":"3827001de363c86c6a22a37f8e35390e4219e81a316c5565ebd4ce89181e8e4e"} Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.652392 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" event={"ID":"0cf7deca-25aa-4516-869a-ad2a20419d3a","Type":"ContainerStarted","Data":"0d16ac8eb733fae58322b0d25da7ba0582004164c537ad23cfbe2aa69f525424"} Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.659868 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6p2pk" event={"ID":"f0d549a7-b237-46de-8dde-d7aba95cfe65","Type":"ContainerStarted","Data":"ee3723be011773b925288457e27ea6a3342f89e0129ec9a39234237792e238b2"} Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.668384 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-pgr62" event={"ID":"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee","Type":"ContainerStarted","Data":"3985626d4c0073f71f868af55f4ae83665d63b16727cd39283a5ab4a453a294d"} Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.686783 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.688444 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74eb-account-create-update-hhzh4" event={"ID":"81d9684c-2d7b-49df-a5c6-a56656a84874","Type":"ContainerStarted","Data":"d7b3767b968920f4c2b24c5037f410fc0f2515fc68b40d168482612b5507bc94"} Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.701236 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e92a-account-create-update-6bjzf" event={"ID":"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20","Type":"ContainerStarted","Data":"c0807268420a67fe4f110c9f1bf5fabeed9719cc45407b2a5897a39fe8c4486a"} Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.701315 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-b7qxz" Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.811920 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b7qxz"] Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.816957 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.847538 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b7qxz"] Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.927011 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-pgr62"] Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.980438 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-cl5cv"] Dec 03 08:59:03 crc kubenswrapper[4573]: E1203 08:59:03.980989 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" containerName="init" Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.981015 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" containerName="init" Dec 03 08:59:03 crc kubenswrapper[4573]: E1203 08:59:03.981045 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" containerName="dnsmasq-dns" Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.981055 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" containerName="dnsmasq-dns" Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.981312 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" containerName="dnsmasq-dns" Dec 03 08:59:03 crc kubenswrapper[4573]: I1203 08:59:03.982379 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.031587 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-cl5cv"] Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.064773 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b728cb5-fb6d-48e7-b170-238f76b6ec51" path="/var/lib/kubelet/pods/4b728cb5-fb6d-48e7-b170-238f76b6ec51/volumes" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.179005 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.179398 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltxmg\" (UniqueName: \"kubernetes.io/projected/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-kube-api-access-ltxmg\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.179822 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.179942 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.179984 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-config\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.281814 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.281863 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-config\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.281920 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.281950 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltxmg\" (UniqueName: \"kubernetes.io/projected/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-kube-api-access-ltxmg\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.281988 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.282985 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.282989 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.283608 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.284220 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-config\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.300534 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltxmg\" (UniqueName: \"kubernetes.io/projected/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-kube-api-access-ltxmg\") pod \"dnsmasq-dns-b8fbc5445-cl5cv\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.353750 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.713347 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1999a198-5df7-40fc-ba0f-5b650b3027a6","Type":"ContainerStarted","Data":"9615a1c34f17e4a51a537dcdc60f046ad2d9b1f3108e8a658ac6fcc60bf7f2a6"} Dec 03 08:59:04 crc kubenswrapper[4573]: I1203 08:59:04.915280 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-cl5cv"] Dec 03 08:59:04 crc kubenswrapper[4573]: W1203 08:59:04.937446 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3e18b14_1f6a_4cb6_af58_398d0e393ffe.slice/crio-104b116b43f367c73c2c1ec234bbcf52f46ca85bdda9360dab38b8e34d889688 WatchSource:0}: Error finding container 104b116b43f367c73c2c1ec234bbcf52f46ca85bdda9360dab38b8e34d889688: Status 404 returned error can't find the container with id 104b116b43f367c73c2c1ec234bbcf52f46ca85bdda9360dab38b8e34d889688 Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.142875 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.215185 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.228628 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.228946 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-gbt9b" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.230163 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.230249 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.244927 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.413794 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/053ecb51-80b3-4888-9384-8438ba293ae8-lock\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.414691 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/053ecb51-80b3-4888-9384-8438ba293ae8-cache\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.414813 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.414927 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvs4z\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-kube-api-access-rvs4z\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.415028 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.517038 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.517228 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/053ecb51-80b3-4888-9384-8438ba293ae8-lock\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.517277 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/053ecb51-80b3-4888-9384-8438ba293ae8-cache\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.517342 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.517371 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvs4z\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-kube-api-access-rvs4z\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: E1203 08:59:05.518226 4573 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 08:59:05 crc kubenswrapper[4573]: E1203 08:59:05.518262 4573 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 08:59:05 crc kubenswrapper[4573]: E1203 08:59:05.518316 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift podName:053ecb51-80b3-4888-9384-8438ba293ae8 nodeName:}" failed. No retries permitted until 2025-12-03 08:59:06.01829474 +0000 UTC m=+1266.586673999 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift") pod "swift-storage-0" (UID: "053ecb51-80b3-4888-9384-8438ba293ae8") : configmap "swift-ring-files" not found Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.518597 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.533381 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/053ecb51-80b3-4888-9384-8438ba293ae8-cache\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.535256 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/053ecb51-80b3-4888-9384-8438ba293ae8-lock\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.552206 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvs4z\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-kube-api-access-rvs4z\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.576158 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.730016 4573 generic.go:334] "Generic (PLEG): container finished" podID="b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" containerID="459cddae57aa1664705a44b30056d23bac71369423a843d2d53e98b25d845c6b" exitCode=0 Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.730360 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-pgr62" event={"ID":"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee","Type":"ContainerDied","Data":"459cddae57aa1664705a44b30056d23bac71369423a843d2d53e98b25d845c6b"} Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.741897 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74eb-account-create-update-hhzh4" event={"ID":"81d9684c-2d7b-49df-a5c6-a56656a84874","Type":"ContainerStarted","Data":"ca2e5f9be83477a0de46d55c12f120cdb471db124cc18e1b801aa299dd9e0563"} Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.750036 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nfl5h" event={"ID":"dd12c3b6-45e8-44dd-b28f-b7c828cb5690","Type":"ContainerStarted","Data":"a8254e75bc483a40ed1b50671feca39018863f80357e2b79c542ea378cfc4d46"} Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.753922 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.776408 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-rrcpn" event={"ID":"6a19cac1-e9f7-4f7d-b356-3263c1b8c446","Type":"ContainerStarted","Data":"79c9f98b9b28da24f113853d9f312623ca3a36b10c35e7b58260e20955137c4a"} Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.802428 4573 generic.go:334] "Generic (PLEG): container finished" podID="0cf7deca-25aa-4516-869a-ad2a20419d3a" containerID="b0ed74d6a2693a5764a1b628123edd07fd98031b2287c41e513ee2fb13d41f45" exitCode=0 Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.802569 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" event={"ID":"0cf7deca-25aa-4516-869a-ad2a20419d3a","Type":"ContainerDied","Data":"b0ed74d6a2693a5764a1b628123edd07fd98031b2287c41e513ee2fb13d41f45"} Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.815771 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-nfl5h" podStartSLOduration=4.815732665 podStartE2EDuration="4.815732665s" podCreationTimestamp="2025-12-03 08:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:05.80270242 +0000 UTC m=+1266.371081689" watchObservedRunningTime="2025-12-03 08:59:05.815732665 +0000 UTC m=+1266.384111924" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.844614 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6p2pk" event={"ID":"f0d549a7-b237-46de-8dde-d7aba95cfe65","Type":"ContainerStarted","Data":"f50555e5e1b886b27f9be5f4290762ef64b328d70615ead4dc38bf3a0cf88cd2"} Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.846117 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-74eb-account-create-update-hhzh4" podStartSLOduration=5.846082156 podStartE2EDuration="5.846082156s" podCreationTimestamp="2025-12-03 08:59:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:05.844675008 +0000 UTC m=+1266.413054277" watchObservedRunningTime="2025-12-03 08:59:05.846082156 +0000 UTC m=+1266.414461415" Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.858920 4573 generic.go:334] "Generic (PLEG): container finished" podID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" containerID="823decdf41f2767a07f9b8a0f91c06403b0695e8cd0d75c5f5350ba36ed7b5c9" exitCode=0 Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.859129 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" event={"ID":"d3e18b14-1f6a-4cb6-af58-398d0e393ffe","Type":"ContainerDied","Data":"823decdf41f2767a07f9b8a0f91c06403b0695e8cd0d75c5f5350ba36ed7b5c9"} Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.859171 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" event={"ID":"d3e18b14-1f6a-4cb6-af58-398d0e393ffe","Type":"ContainerStarted","Data":"104b116b43f367c73c2c1ec234bbcf52f46ca85bdda9360dab38b8e34d889688"} Dec 03 08:59:05 crc kubenswrapper[4573]: I1203 08:59:05.874826 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e92a-account-create-update-6bjzf" event={"ID":"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20","Type":"ContainerStarted","Data":"c19b4d5e1262e2b1bb5fd37ae2944d1b0fb9eeee59d7cf0ec7e54a49bb92c0c6"} Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.001910 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-6p2pk" podStartSLOduration=6.001885037 podStartE2EDuration="6.001885037s" podCreationTimestamp="2025-12-03 08:59:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:05.903532047 +0000 UTC m=+1266.471911306" watchObservedRunningTime="2025-12-03 08:59:06.001885037 +0000 UTC m=+1266.570264296" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.022545 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-e92a-account-create-update-6bjzf" podStartSLOduration=5.022525751 podStartE2EDuration="5.022525751s" podCreationTimestamp="2025-12-03 08:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:06.003348567 +0000 UTC m=+1266.571727826" watchObservedRunningTime="2025-12-03 08:59:06.022525751 +0000 UTC m=+1266.590905011" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.043858 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:06 crc kubenswrapper[4573]: E1203 08:59:06.045191 4573 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 08:59:06 crc kubenswrapper[4573]: E1203 08:59:06.045218 4573 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 08:59:06 crc kubenswrapper[4573]: E1203 08:59:06.045275 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift podName:053ecb51-80b3-4888-9384-8438ba293ae8 nodeName:}" failed. No retries permitted until 2025-12-03 08:59:07.045254253 +0000 UTC m=+1267.613633512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift") pod "swift-storage-0" (UID: "053ecb51-80b3-4888-9384-8438ba293ae8") : configmap "swift-ring-files" not found Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.074651 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.095761 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-rrcpn" podStartSLOduration=5.095732734 podStartE2EDuration="5.095732734s" podCreationTimestamp="2025-12-03 08:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:06.071695557 +0000 UTC m=+1266.640074816" watchObservedRunningTime="2025-12-03 08:59:06.095732734 +0000 UTC m=+1266.664111993" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.455426 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.573497 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx4hf\" (UniqueName: \"kubernetes.io/projected/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-kube-api-access-jx4hf\") pod \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.573616 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-config\") pod \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.573695 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-dns-svc\") pod \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.573723 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-nb\") pod \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.573782 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-sb\") pod \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\" (UID: \"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.586308 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-kube-api-access-jx4hf" (OuterVolumeSpecName: "kube-api-access-jx4hf") pod "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" (UID: "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee"). InnerVolumeSpecName "kube-api-access-jx4hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.614364 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-xd2w2"] Dec 03 08:59:06 crc kubenswrapper[4573]: E1203 08:59:06.620330 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" containerName="init" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.620383 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" containerName="init" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.624287 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" containerName="init" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.626643 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.637171 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xd2w2"] Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.665610 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.675887 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" (UID: "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.676206 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx4hf\" (UniqueName: \"kubernetes.io/projected/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-kube-api-access-jx4hf\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.676235 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.689037 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-config" (OuterVolumeSpecName: "config") pod "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" (UID: "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.713412 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" (UID: "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.734148 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" (UID: "b5fa77a1-fea1-4f30-8076-ad7e4bab23ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.745226 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-3c12-account-create-update-ntcsm"] Dec 03 08:59:06 crc kubenswrapper[4573]: E1203 08:59:06.745843 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf7deca-25aa-4516-869a-ad2a20419d3a" containerName="init" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.745870 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf7deca-25aa-4516-869a-ad2a20419d3a" containerName="init" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.746180 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf7deca-25aa-4516-869a-ad2a20419d3a" containerName="init" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.747213 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.750149 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.787607 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-ovsdbserver-nb\") pod \"0cf7deca-25aa-4516-869a-ad2a20419d3a\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.787779 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-config\") pod \"0cf7deca-25aa-4516-869a-ad2a20419d3a\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.787856 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7dt9\" (UniqueName: \"kubernetes.io/projected/0cf7deca-25aa-4516-869a-ad2a20419d3a-kube-api-access-m7dt9\") pod \"0cf7deca-25aa-4516-869a-ad2a20419d3a\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.787936 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-dns-svc\") pod \"0cf7deca-25aa-4516-869a-ad2a20419d3a\" (UID: \"0cf7deca-25aa-4516-869a-ad2a20419d3a\") " Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.789873 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62ff626-7df4-4756-ab26-a2d74f71bb57-operator-scripts\") pod \"glance-db-create-xd2w2\" (UID: \"d62ff626-7df4-4756-ab26-a2d74f71bb57\") " pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.790878 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7h5t\" (UniqueName: \"kubernetes.io/projected/d62ff626-7df4-4756-ab26-a2d74f71bb57-kube-api-access-n7h5t\") pod \"glance-db-create-xd2w2\" (UID: \"d62ff626-7df4-4756-ab26-a2d74f71bb57\") " pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.791141 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.791157 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.791167 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.792407 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3c12-account-create-update-ntcsm"] Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.800200 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cf7deca-25aa-4516-869a-ad2a20419d3a-kube-api-access-m7dt9" (OuterVolumeSpecName: "kube-api-access-m7dt9") pod "0cf7deca-25aa-4516-869a-ad2a20419d3a" (UID: "0cf7deca-25aa-4516-869a-ad2a20419d3a"). InnerVolumeSpecName "kube-api-access-m7dt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.826641 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0cf7deca-25aa-4516-869a-ad2a20419d3a" (UID: "0cf7deca-25aa-4516-869a-ad2a20419d3a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.855617 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-config" (OuterVolumeSpecName: "config") pod "0cf7deca-25aa-4516-869a-ad2a20419d3a" (UID: "0cf7deca-25aa-4516-869a-ad2a20419d3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.896765 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62ff626-7df4-4756-ab26-a2d74f71bb57-operator-scripts\") pod \"glance-db-create-xd2w2\" (UID: \"d62ff626-7df4-4756-ab26-a2d74f71bb57\") " pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.896886 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxqx2\" (UniqueName: \"kubernetes.io/projected/ffbef148-b74e-4d49-ab26-1c1e9808074f-kube-api-access-gxqx2\") pod \"glance-3c12-account-create-update-ntcsm\" (UID: \"ffbef148-b74e-4d49-ab26-1c1e9808074f\") " pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.897023 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffbef148-b74e-4d49-ab26-1c1e9808074f-operator-scripts\") pod \"glance-3c12-account-create-update-ntcsm\" (UID: \"ffbef148-b74e-4d49-ab26-1c1e9808074f\") " pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.897248 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7h5t\" (UniqueName: \"kubernetes.io/projected/d62ff626-7df4-4756-ab26-a2d74f71bb57-kube-api-access-n7h5t\") pod \"glance-db-create-xd2w2\" (UID: \"d62ff626-7df4-4756-ab26-a2d74f71bb57\") " pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.897379 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.897398 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.897422 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7dt9\" (UniqueName: \"kubernetes.io/projected/0cf7deca-25aa-4516-869a-ad2a20419d3a-kube-api-access-m7dt9\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.899880 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62ff626-7df4-4756-ab26-a2d74f71bb57-operator-scripts\") pod \"glance-db-create-xd2w2\" (UID: \"d62ff626-7df4-4756-ab26-a2d74f71bb57\") " pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.925676 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0cf7deca-25aa-4516-869a-ad2a20419d3a" (UID: "0cf7deca-25aa-4516-869a-ad2a20419d3a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.928569 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7h5t\" (UniqueName: \"kubernetes.io/projected/d62ff626-7df4-4756-ab26-a2d74f71bb57-kube-api-access-n7h5t\") pod \"glance-db-create-xd2w2\" (UID: \"d62ff626-7df4-4756-ab26-a2d74f71bb57\") " pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.930182 4573 generic.go:334] "Generic (PLEG): container finished" podID="81d9684c-2d7b-49df-a5c6-a56656a84874" containerID="ca2e5f9be83477a0de46d55c12f120cdb471db124cc18e1b801aa299dd9e0563" exitCode=0 Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.930328 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74eb-account-create-update-hhzh4" event={"ID":"81d9684c-2d7b-49df-a5c6-a56656a84874","Type":"ContainerDied","Data":"ca2e5f9be83477a0de46d55c12f120cdb471db124cc18e1b801aa299dd9e0563"} Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.938219 4573 generic.go:334] "Generic (PLEG): container finished" podID="dd12c3b6-45e8-44dd-b28f-b7c828cb5690" containerID="a8254e75bc483a40ed1b50671feca39018863f80357e2b79c542ea378cfc4d46" exitCode=0 Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.938307 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nfl5h" event={"ID":"dd12c3b6-45e8-44dd-b28f-b7c828cb5690","Type":"ContainerDied","Data":"a8254e75bc483a40ed1b50671feca39018863f80357e2b79c542ea378cfc4d46"} Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.939750 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" event={"ID":"0cf7deca-25aa-4516-869a-ad2a20419d3a","Type":"ContainerDied","Data":"0d16ac8eb733fae58322b0d25da7ba0582004164c537ad23cfbe2aa69f525424"} Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.939917 4573 scope.go:117] "RemoveContainer" containerID="b0ed74d6a2693a5764a1b628123edd07fd98031b2287c41e513ee2fb13d41f45" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.940255 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-dgch8" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.944314 4573 generic.go:334] "Generic (PLEG): container finished" podID="f0d549a7-b237-46de-8dde-d7aba95cfe65" containerID="f50555e5e1b886b27f9be5f4290762ef64b328d70615ead4dc38bf3a0cf88cd2" exitCode=0 Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.944423 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6p2pk" event={"ID":"f0d549a7-b237-46de-8dde-d7aba95cfe65","Type":"ContainerDied","Data":"f50555e5e1b886b27f9be5f4290762ef64b328d70615ead4dc38bf3a0cf88cd2"} Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.946932 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" event={"ID":"d3e18b14-1f6a-4cb6-af58-398d0e393ffe","Type":"ContainerStarted","Data":"6ef8e5934c94595f86df204e3ac8183e89565374ccead7c035e8695fe6a1b12c"} Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.948230 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.952673 4573 generic.go:334] "Generic (PLEG): container finished" podID="e78a67e6-fbd8-4c2b-bc2d-148de6b27c20" containerID="c19b4d5e1262e2b1bb5fd37ae2944d1b0fb9eeee59d7cf0ec7e54a49bb92c0c6" exitCode=0 Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.952772 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e92a-account-create-update-6bjzf" event={"ID":"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20","Type":"ContainerDied","Data":"c19b4d5e1262e2b1bb5fd37ae2944d1b0fb9eeee59d7cf0ec7e54a49bb92c0c6"} Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.978654 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-pgr62" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.978405 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-pgr62" event={"ID":"b5fa77a1-fea1-4f30-8076-ad7e4bab23ee","Type":"ContainerDied","Data":"3985626d4c0073f71f868af55f4ae83665d63b16727cd39283a5ab4a453a294d"} Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.980077 4573 scope.go:117] "RemoveContainer" containerID="459cddae57aa1664705a44b30056d23bac71369423a843d2d53e98b25d845c6b" Dec 03 08:59:06 crc kubenswrapper[4573]: I1203 08:59:06.999310 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxqx2\" (UniqueName: \"kubernetes.io/projected/ffbef148-b74e-4d49-ab26-1c1e9808074f-kube-api-access-gxqx2\") pod \"glance-3c12-account-create-update-ntcsm\" (UID: \"ffbef148-b74e-4d49-ab26-1c1e9808074f\") " pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.000082 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffbef148-b74e-4d49-ab26-1c1e9808074f-operator-scripts\") pod \"glance-3c12-account-create-update-ntcsm\" (UID: \"ffbef148-b74e-4d49-ab26-1c1e9808074f\") " pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.000312 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cf7deca-25aa-4516-869a-ad2a20419d3a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.005300 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffbef148-b74e-4d49-ab26-1c1e9808074f-operator-scripts\") pod \"glance-3c12-account-create-update-ntcsm\" (UID: \"ffbef148-b74e-4d49-ab26-1c1e9808074f\") " pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.033374 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxqx2\" (UniqueName: \"kubernetes.io/projected/ffbef148-b74e-4d49-ab26-1c1e9808074f-kube-api-access-gxqx2\") pod \"glance-3c12-account-create-update-ntcsm\" (UID: \"ffbef148-b74e-4d49-ab26-1c1e9808074f\") " pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.033852 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" podStartSLOduration=4.033825061 podStartE2EDuration="4.033825061s" podCreationTimestamp="2025-12-03 08:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:07.020550259 +0000 UTC m=+1267.588929528" watchObservedRunningTime="2025-12-03 08:59:07.033825061 +0000 UTC m=+1267.602204320" Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.102495 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:07 crc kubenswrapper[4573]: E1203 08:59:07.104498 4573 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 08:59:07 crc kubenswrapper[4573]: E1203 08:59:07.104669 4573 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 08:59:07 crc kubenswrapper[4573]: E1203 08:59:07.104849 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift podName:053ecb51-80b3-4888-9384-8438ba293ae8 nodeName:}" failed. No retries permitted until 2025-12-03 08:59:09.104816684 +0000 UTC m=+1269.673196103 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift") pod "swift-storage-0" (UID: "053ecb51-80b3-4888-9384-8438ba293ae8") : configmap "swift-ring-files" not found Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.128730 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-pgr62"] Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.129168 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.137368 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-pgr62"] Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.167924 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.189453 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-dgch8"] Dec 03 08:59:07 crc kubenswrapper[4573]: I1203 08:59:07.199913 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-dgch8"] Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.060702 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cf7deca-25aa-4516-869a-ad2a20419d3a" path="/var/lib/kubelet/pods/0cf7deca-25aa-4516-869a-ad2a20419d3a/volumes" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.064410 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5fa77a1-fea1-4f30-8076-ad7e4bab23ee" path="/var/lib/kubelet/pods/b5fa77a1-fea1-4f30-8076-ad7e4bab23ee/volumes" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.221521 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xd2w2"] Dec 03 08:59:08 crc kubenswrapper[4573]: W1203 08:59:08.233844 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd62ff626_7df4_4756_ab26_a2d74f71bb57.slice/crio-d3fc84a260fad75cb4eb94fbfe947a7fd6a2e8586d4fac25c945ddf1b0a47299 WatchSource:0}: Error finding container d3fc84a260fad75cb4eb94fbfe947a7fd6a2e8586d4fac25c945ddf1b0a47299: Status 404 returned error can't find the container with id d3fc84a260fad75cb4eb94fbfe947a7fd6a2e8586d4fac25c945ddf1b0a47299 Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.358914 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3c12-account-create-update-ntcsm"] Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.511610 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.573144 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-operator-scripts\") pod \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\" (UID: \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\") " Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.573374 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpnbq\" (UniqueName: \"kubernetes.io/projected/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-kube-api-access-lpnbq\") pod \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\" (UID: \"dd12c3b6-45e8-44dd-b28f-b7c828cb5690\") " Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.574880 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd12c3b6-45e8-44dd-b28f-b7c828cb5690" (UID: "dd12c3b6-45e8-44dd-b28f-b7c828cb5690"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.660320 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-kube-api-access-lpnbq" (OuterVolumeSpecName: "kube-api-access-lpnbq") pod "dd12c3b6-45e8-44dd-b28f-b7c828cb5690" (UID: "dd12c3b6-45e8-44dd-b28f-b7c828cb5690"). InnerVolumeSpecName "kube-api-access-lpnbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.660381 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.675407 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d9684c-2d7b-49df-a5c6-a56656a84874-operator-scripts\") pod \"81d9684c-2d7b-49df-a5c6-a56656a84874\" (UID: \"81d9684c-2d7b-49df-a5c6-a56656a84874\") " Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.675538 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44vhh\" (UniqueName: \"kubernetes.io/projected/81d9684c-2d7b-49df-a5c6-a56656a84874-kube-api-access-44vhh\") pod \"81d9684c-2d7b-49df-a5c6-a56656a84874\" (UID: \"81d9684c-2d7b-49df-a5c6-a56656a84874\") " Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.676110 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81d9684c-2d7b-49df-a5c6-a56656a84874-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81d9684c-2d7b-49df-a5c6-a56656a84874" (UID: "81d9684c-2d7b-49df-a5c6-a56656a84874"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.676252 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.676279 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpnbq\" (UniqueName: \"kubernetes.io/projected/dd12c3b6-45e8-44dd-b28f-b7c828cb5690-kube-api-access-lpnbq\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.676294 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81d9684c-2d7b-49df-a5c6-a56656a84874-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.698093 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81d9684c-2d7b-49df-a5c6-a56656a84874-kube-api-access-44vhh" (OuterVolumeSpecName: "kube-api-access-44vhh") pod "81d9684c-2d7b-49df-a5c6-a56656a84874" (UID: "81d9684c-2d7b-49df-a5c6-a56656a84874"). InnerVolumeSpecName "kube-api-access-44vhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.780750 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44vhh\" (UniqueName: \"kubernetes.io/projected/81d9684c-2d7b-49df-a5c6-a56656a84874-kube-api-access-44vhh\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.897427 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.904366 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.965244 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-lzbxl"] Dec 03 08:59:08 crc kubenswrapper[4573]: E1203 08:59:08.967289 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81d9684c-2d7b-49df-a5c6-a56656a84874" containerName="mariadb-account-create-update" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.967317 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="81d9684c-2d7b-49df-a5c6-a56656a84874" containerName="mariadb-account-create-update" Dec 03 08:59:08 crc kubenswrapper[4573]: E1203 08:59:08.967331 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd12c3b6-45e8-44dd-b28f-b7c828cb5690" containerName="mariadb-database-create" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.967343 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd12c3b6-45e8-44dd-b28f-b7c828cb5690" containerName="mariadb-database-create" Dec 03 08:59:08 crc kubenswrapper[4573]: E1203 08:59:08.967365 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78a67e6-fbd8-4c2b-bc2d-148de6b27c20" containerName="mariadb-account-create-update" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.967374 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78a67e6-fbd8-4c2b-bc2d-148de6b27c20" containerName="mariadb-account-create-update" Dec 03 08:59:08 crc kubenswrapper[4573]: E1203 08:59:08.967403 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d549a7-b237-46de-8dde-d7aba95cfe65" containerName="mariadb-database-create" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.967410 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d549a7-b237-46de-8dde-d7aba95cfe65" containerName="mariadb-database-create" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.968863 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="81d9684c-2d7b-49df-a5c6-a56656a84874" containerName="mariadb-account-create-update" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.968926 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd12c3b6-45e8-44dd-b28f-b7c828cb5690" containerName="mariadb-database-create" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.968946 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78a67e6-fbd8-4c2b-bc2d-148de6b27c20" containerName="mariadb-account-create-update" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.968961 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d549a7-b237-46de-8dde-d7aba95cfe65" containerName="mariadb-database-create" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.971638 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.975027 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.975942 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.976176 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.989636 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-swiftconf\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.989965 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-dispersionconf\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.990108 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ceed462-6691-4229-b764-b25eb42bbf86-etc-swift\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.990187 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-scripts\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:08 crc kubenswrapper[4573]: I1203 08:59:08.992902 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-lzbxl"] Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:08.997738 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj6zb\" (UniqueName: \"kubernetes.io/projected/5ceed462-6691-4229-b764-b25eb42bbf86-kube-api-access-pj6zb\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:08.997837 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-ring-data-devices\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:08.997859 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-combined-ca-bundle\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.035702 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xd2w2" event={"ID":"d62ff626-7df4-4756-ab26-a2d74f71bb57","Type":"ContainerStarted","Data":"ffc97e4137bae2488e3f6389c3fb5e75e5948c598147f5d1e8772f55cb0acb3d"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.035878 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xd2w2" event={"ID":"d62ff626-7df4-4756-ab26-a2d74f71bb57","Type":"ContainerStarted","Data":"d3fc84a260fad75cb4eb94fbfe947a7fd6a2e8586d4fac25c945ddf1b0a47299"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.060824 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nfl5h" event={"ID":"dd12c3b6-45e8-44dd-b28f-b7c828cb5690","Type":"ContainerDied","Data":"8029c068fbce6aa6701f1dc03f3f350fab15e5a0a82d97c7126282f51c95b531"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.060889 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8029c068fbce6aa6701f1dc03f3f350fab15e5a0a82d97c7126282f51c95b531" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.061095 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nfl5h" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.088329 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3c12-account-create-update-ntcsm" event={"ID":"ffbef148-b74e-4d49-ab26-1c1e9808074f","Type":"ContainerStarted","Data":"07f699b6a8d9d47e724de128143d562bd9b9db501e43b50f272b4a4b256ef4db"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.088399 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3c12-account-create-update-ntcsm" event={"ID":"ffbef148-b74e-4d49-ab26-1c1e9808074f","Type":"ContainerStarted","Data":"95a5bd9fe4cb805c6b4b57a73d1343b62ee9b45b0b8a6111234b27d8f2d4b642"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.107270 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62wtm\" (UniqueName: \"kubernetes.io/projected/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-kube-api-access-62wtm\") pod \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\" (UID: \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\") " Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.107851 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-operator-scripts\") pod \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\" (UID: \"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20\") " Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108194 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0d549a7-b237-46de-8dde-d7aba95cfe65-operator-scripts\") pod \"f0d549a7-b237-46de-8dde-d7aba95cfe65\" (UID: \"f0d549a7-b237-46de-8dde-d7aba95cfe65\") " Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108223 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cblvq\" (UniqueName: \"kubernetes.io/projected/f0d549a7-b237-46de-8dde-d7aba95cfe65-kube-api-access-cblvq\") pod \"f0d549a7-b237-46de-8dde-d7aba95cfe65\" (UID: \"f0d549a7-b237-46de-8dde-d7aba95cfe65\") " Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108646 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-dispersionconf\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108691 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ceed462-6691-4229-b764-b25eb42bbf86-etc-swift\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108716 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-scripts\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108793 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108825 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj6zb\" (UniqueName: \"kubernetes.io/projected/5ceed462-6691-4229-b764-b25eb42bbf86-kube-api-access-pj6zb\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108848 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-ring-data-devices\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108867 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-combined-ca-bundle\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.108920 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-swiftconf\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.109070 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e78a67e6-fbd8-4c2b-bc2d-148de6b27c20" (UID: "e78a67e6-fbd8-4c2b-bc2d-148de6b27c20"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.109438 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0d549a7-b237-46de-8dde-d7aba95cfe65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0d549a7-b237-46de-8dde-d7aba95cfe65" (UID: "f0d549a7-b237-46de-8dde-d7aba95cfe65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.111153 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1999a198-5df7-40fc-ba0f-5b650b3027a6","Type":"ContainerStarted","Data":"176260df37a569a76a222d75d9037b3f1f8816c3ad9e9739e927688a3db8aa9d"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.110873 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-scripts\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.113679 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-ring-data-devices\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: E1203 08:59:09.114608 4573 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 08:59:09 crc kubenswrapper[4573]: E1203 08:59:09.114626 4573 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 08:59:09 crc kubenswrapper[4573]: E1203 08:59:09.114673 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift podName:053ecb51-80b3-4888-9384-8438ba293ae8 nodeName:}" failed. No retries permitted until 2025-12-03 08:59:13.114656574 +0000 UTC m=+1273.683035833 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift") pod "swift-storage-0" (UID: "053ecb51-80b3-4888-9384-8438ba293ae8") : configmap "swift-ring-files" not found Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.121179 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.121240 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0d549a7-b237-46de-8dde-d7aba95cfe65-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.111202 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1999a198-5df7-40fc-ba0f-5b650b3027a6","Type":"ContainerStarted","Data":"e46b34c1020eea36ffd5a4012396858473c6d53a0ec696666d197f770a58cc88"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.126948 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ceed462-6691-4229-b764-b25eb42bbf86-etc-swift\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.127570 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-combined-ca-bundle\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.127956 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.128965 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6p2pk" event={"ID":"f0d549a7-b237-46de-8dde-d7aba95cfe65","Type":"ContainerDied","Data":"ee3723be011773b925288457e27ea6a3342f89e0129ec9a39234237792e238b2"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.129006 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee3723be011773b925288457e27ea6a3342f89e0129ec9a39234237792e238b2" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.128996 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-swiftconf\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.129009 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6p2pk" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.130475 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0d549a7-b237-46de-8dde-d7aba95cfe65-kube-api-access-cblvq" (OuterVolumeSpecName: "kube-api-access-cblvq") pod "f0d549a7-b237-46de-8dde-d7aba95cfe65" (UID: "f0d549a7-b237-46de-8dde-d7aba95cfe65"). InnerVolumeSpecName "kube-api-access-cblvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.130590 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-kube-api-access-62wtm" (OuterVolumeSpecName: "kube-api-access-62wtm") pod "e78a67e6-fbd8-4c2b-bc2d-148de6b27c20" (UID: "e78a67e6-fbd8-4c2b-bc2d-148de6b27c20"). InnerVolumeSpecName "kube-api-access-62wtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.142098 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-dispersionconf\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.144460 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e92a-account-create-update-6bjzf" event={"ID":"e78a67e6-fbd8-4c2b-bc2d-148de6b27c20","Type":"ContainerDied","Data":"c0807268420a67fe4f110c9f1bf5fabeed9719cc45407b2a5897a39fe8c4486a"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.144502 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0807268420a67fe4f110c9f1bf5fabeed9719cc45407b2a5897a39fe8c4486a" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.144627 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e92a-account-create-update-6bjzf" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.149065 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-xd2w2" podStartSLOduration=3.149025434 podStartE2EDuration="3.149025434s" podCreationTimestamp="2025-12-03 08:59:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:09.088956221 +0000 UTC m=+1269.657335480" watchObservedRunningTime="2025-12-03 08:59:09.149025434 +0000 UTC m=+1269.717404693" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.151026 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-74eb-account-create-update-hhzh4" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.154421 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-74eb-account-create-update-hhzh4" event={"ID":"81d9684c-2d7b-49df-a5c6-a56656a84874","Type":"ContainerDied","Data":"d7b3767b968920f4c2b24c5037f410fc0f2515fc68b40d168482612b5507bc94"} Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.154515 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7b3767b968920f4c2b24c5037f410fc0f2515fc68b40d168482612b5507bc94" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.165339 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-3c12-account-create-update-ntcsm" podStartSLOduration=3.165303249 podStartE2EDuration="3.165303249s" podCreationTimestamp="2025-12-03 08:59:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:09.127516086 +0000 UTC m=+1269.695895355" watchObservedRunningTime="2025-12-03 08:59:09.165303249 +0000 UTC m=+1269.733682508" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.180373 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.307321819 podStartE2EDuration="7.18034855s" podCreationTimestamp="2025-12-03 08:59:02 +0000 UTC" firstStartedPulling="2025-12-03 08:59:03.767628478 +0000 UTC m=+1264.336007727" lastFinishedPulling="2025-12-03 08:59:07.640655199 +0000 UTC m=+1268.209034458" observedRunningTime="2025-12-03 08:59:09.178214873 +0000 UTC m=+1269.746594132" watchObservedRunningTime="2025-12-03 08:59:09.18034855 +0000 UTC m=+1269.748727809" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.182629 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj6zb\" (UniqueName: \"kubernetes.io/projected/5ceed462-6691-4229-b764-b25eb42bbf86-kube-api-access-pj6zb\") pod \"swift-ring-rebalance-lzbxl\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.227832 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cblvq\" (UniqueName: \"kubernetes.io/projected/f0d549a7-b237-46de-8dde-d7aba95cfe65-kube-api-access-cblvq\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.227874 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62wtm\" (UniqueName: \"kubernetes.io/projected/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20-kube-api-access-62wtm\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.315716 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:09 crc kubenswrapper[4573]: I1203 08:59:09.971204 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-lzbxl"] Dec 03 08:59:10 crc kubenswrapper[4573]: I1203 08:59:10.161795 4573 generic.go:334] "Generic (PLEG): container finished" podID="ffbef148-b74e-4d49-ab26-1c1e9808074f" containerID="07f699b6a8d9d47e724de128143d562bd9b9db501e43b50f272b4a4b256ef4db" exitCode=0 Dec 03 08:59:10 crc kubenswrapper[4573]: I1203 08:59:10.161905 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3c12-account-create-update-ntcsm" event={"ID":"ffbef148-b74e-4d49-ab26-1c1e9808074f","Type":"ContainerDied","Data":"07f699b6a8d9d47e724de128143d562bd9b9db501e43b50f272b4a4b256ef4db"} Dec 03 08:59:10 crc kubenswrapper[4573]: I1203 08:59:10.163399 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-lzbxl" event={"ID":"5ceed462-6691-4229-b764-b25eb42bbf86","Type":"ContainerStarted","Data":"e477364d9e48d69d2b0ea21df0df222d2f40fbab44e1c9af982cc1f56359e858"} Dec 03 08:59:10 crc kubenswrapper[4573]: I1203 08:59:10.168039 4573 generic.go:334] "Generic (PLEG): container finished" podID="d62ff626-7df4-4756-ab26-a2d74f71bb57" containerID="ffc97e4137bae2488e3f6389c3fb5e75e5948c598147f5d1e8772f55cb0acb3d" exitCode=0 Dec 03 08:59:10 crc kubenswrapper[4573]: I1203 08:59:10.168585 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xd2w2" event={"ID":"d62ff626-7df4-4756-ab26-a2d74f71bb57","Type":"ContainerDied","Data":"ffc97e4137bae2488e3f6389c3fb5e75e5948c598147f5d1e8772f55cb0acb3d"} Dec 03 08:59:11 crc kubenswrapper[4573]: I1203 08:59:11.821268 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:11 crc kubenswrapper[4573]: I1203 08:59:11.830080 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.004647 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62ff626-7df4-4756-ab26-a2d74f71bb57-operator-scripts\") pod \"d62ff626-7df4-4756-ab26-a2d74f71bb57\" (UID: \"d62ff626-7df4-4756-ab26-a2d74f71bb57\") " Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.004775 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxqx2\" (UniqueName: \"kubernetes.io/projected/ffbef148-b74e-4d49-ab26-1c1e9808074f-kube-api-access-gxqx2\") pod \"ffbef148-b74e-4d49-ab26-1c1e9808074f\" (UID: \"ffbef148-b74e-4d49-ab26-1c1e9808074f\") " Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.004808 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7h5t\" (UniqueName: \"kubernetes.io/projected/d62ff626-7df4-4756-ab26-a2d74f71bb57-kube-api-access-n7h5t\") pod \"d62ff626-7df4-4756-ab26-a2d74f71bb57\" (UID: \"d62ff626-7df4-4756-ab26-a2d74f71bb57\") " Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.004958 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffbef148-b74e-4d49-ab26-1c1e9808074f-operator-scripts\") pod \"ffbef148-b74e-4d49-ab26-1c1e9808074f\" (UID: \"ffbef148-b74e-4d49-ab26-1c1e9808074f\") " Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.005652 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d62ff626-7df4-4756-ab26-a2d74f71bb57-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d62ff626-7df4-4756-ab26-a2d74f71bb57" (UID: "d62ff626-7df4-4756-ab26-a2d74f71bb57"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.005850 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d62ff626-7df4-4756-ab26-a2d74f71bb57-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.006240 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffbef148-b74e-4d49-ab26-1c1e9808074f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ffbef148-b74e-4d49-ab26-1c1e9808074f" (UID: "ffbef148-b74e-4d49-ab26-1c1e9808074f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.017352 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d62ff626-7df4-4756-ab26-a2d74f71bb57-kube-api-access-n7h5t" (OuterVolumeSpecName: "kube-api-access-n7h5t") pod "d62ff626-7df4-4756-ab26-a2d74f71bb57" (UID: "d62ff626-7df4-4756-ab26-a2d74f71bb57"). InnerVolumeSpecName "kube-api-access-n7h5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.017424 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffbef148-b74e-4d49-ab26-1c1e9808074f-kube-api-access-gxqx2" (OuterVolumeSpecName: "kube-api-access-gxqx2") pod "ffbef148-b74e-4d49-ab26-1c1e9808074f" (UID: "ffbef148-b74e-4d49-ab26-1c1e9808074f"). InnerVolumeSpecName "kube-api-access-gxqx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.113504 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxqx2\" (UniqueName: \"kubernetes.io/projected/ffbef148-b74e-4d49-ab26-1c1e9808074f-kube-api-access-gxqx2\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.115362 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7h5t\" (UniqueName: \"kubernetes.io/projected/d62ff626-7df4-4756-ab26-a2d74f71bb57-kube-api-access-n7h5t\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.115491 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ffbef148-b74e-4d49-ab26-1c1e9808074f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.204722 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xd2w2" event={"ID":"d62ff626-7df4-4756-ab26-a2d74f71bb57","Type":"ContainerDied","Data":"d3fc84a260fad75cb4eb94fbfe947a7fd6a2e8586d4fac25c945ddf1b0a47299"} Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.204781 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3fc84a260fad75cb4eb94fbfe947a7fd6a2e8586d4fac25c945ddf1b0a47299" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.204869 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xd2w2" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.208961 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3c12-account-create-update-ntcsm" event={"ID":"ffbef148-b74e-4d49-ab26-1c1e9808074f","Type":"ContainerDied","Data":"95a5bd9fe4cb805c6b4b57a73d1343b62ee9b45b0b8a6111234b27d8f2d4b642"} Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.208991 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95a5bd9fe4cb805c6b4b57a73d1343b62ee9b45b0b8a6111234b27d8f2d4b642" Dec 03 08:59:12 crc kubenswrapper[4573]: I1203 08:59:12.209033 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3c12-account-create-update-ntcsm" Dec 03 08:59:13 crc kubenswrapper[4573]: I1203 08:59:13.138086 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:13 crc kubenswrapper[4573]: E1203 08:59:13.138491 4573 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 08:59:13 crc kubenswrapper[4573]: E1203 08:59:13.138517 4573 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 08:59:13 crc kubenswrapper[4573]: E1203 08:59:13.138600 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift podName:053ecb51-80b3-4888-9384-8438ba293ae8 nodeName:}" failed. No retries permitted until 2025-12-03 08:59:21.138575651 +0000 UTC m=+1281.706954910 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift") pod "swift-storage-0" (UID: "053ecb51-80b3-4888-9384-8438ba293ae8") : configmap "swift-ring-files" not found Dec 03 08:59:14 crc kubenswrapper[4573]: I1203 08:59:14.357925 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 08:59:14 crc kubenswrapper[4573]: I1203 08:59:14.442763 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-h42zx"] Dec 03 08:59:14 crc kubenswrapper[4573]: I1203 08:59:14.443794 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" podUID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" containerName="dnsmasq-dns" containerID="cri-o://5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5" gracePeriod=10 Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.418191 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.419692 4573 generic.go:334] "Generic (PLEG): container finished" podID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" containerID="5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5" exitCode=0 Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.419739 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" event={"ID":"12f89059-a1f2-45fa-be11-e4bdc6b6f936","Type":"ContainerDied","Data":"5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5"} Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.419777 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" event={"ID":"12f89059-a1f2-45fa-be11-e4bdc6b6f936","Type":"ContainerDied","Data":"01276902add22aa019d40c1a2c57cef64747879bcb25b304d2961074dd37f51f"} Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.419798 4573 scope.go:117] "RemoveContainer" containerID="5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.427205 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-config\") pod \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.447558 4573 scope.go:117] "RemoveContainer" containerID="db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.485216 4573 scope.go:117] "RemoveContainer" containerID="5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5" Dec 03 08:59:16 crc kubenswrapper[4573]: E1203 08:59:16.486587 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5\": container with ID starting with 5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5 not found: ID does not exist" containerID="5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.486641 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5"} err="failed to get container status \"5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5\": rpc error: code = NotFound desc = could not find container \"5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5\": container with ID starting with 5475aa93581ee232c495fe6fab9bb84952a5d87afc86cbe893def9ccf19566f5 not found: ID does not exist" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.486691 4573 scope.go:117] "RemoveContainer" containerID="db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0" Dec 03 08:59:16 crc kubenswrapper[4573]: E1203 08:59:16.487085 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0\": container with ID starting with db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0 not found: ID does not exist" containerID="db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.487110 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0"} err="failed to get container status \"db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0\": rpc error: code = NotFound desc = could not find container \"db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0\": container with ID starting with db4bf8e487e8e860355ed620c5ca5d64b4d8f01fff5f9dbec3ddda8a8470fab0 not found: ID does not exist" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.515202 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-config" (OuterVolumeSpecName: "config") pod "12f89059-a1f2-45fa-be11-e4bdc6b6f936" (UID: "12f89059-a1f2-45fa-be11-e4bdc6b6f936"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.528777 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l29p\" (UniqueName: \"kubernetes.io/projected/12f89059-a1f2-45fa-be11-e4bdc6b6f936-kube-api-access-7l29p\") pod \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.529115 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-dns-svc\") pod \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\" (UID: \"12f89059-a1f2-45fa-be11-e4bdc6b6f936\") " Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.529590 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-config\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.541769 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12f89059-a1f2-45fa-be11-e4bdc6b6f936-kube-api-access-7l29p" (OuterVolumeSpecName: "kube-api-access-7l29p") pod "12f89059-a1f2-45fa-be11-e4bdc6b6f936" (UID: "12f89059-a1f2-45fa-be11-e4bdc6b6f936"). InnerVolumeSpecName "kube-api-access-7l29p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.591206 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "12f89059-a1f2-45fa-be11-e4bdc6b6f936" (UID: "12f89059-a1f2-45fa-be11-e4bdc6b6f936"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.631195 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/12f89059-a1f2-45fa-be11-e4bdc6b6f936-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.631247 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l29p\" (UniqueName: \"kubernetes.io/projected/12f89059-a1f2-45fa-be11-e4bdc6b6f936-kube-api-access-7l29p\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.971007 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-8t69f"] Dec 03 08:59:16 crc kubenswrapper[4573]: E1203 08:59:16.971639 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" containerName="init" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.971669 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" containerName="init" Dec 03 08:59:16 crc kubenswrapper[4573]: E1203 08:59:16.971685 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62ff626-7df4-4756-ab26-a2d74f71bb57" containerName="mariadb-database-create" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.971695 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62ff626-7df4-4756-ab26-a2d74f71bb57" containerName="mariadb-database-create" Dec 03 08:59:16 crc kubenswrapper[4573]: E1203 08:59:16.971718 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffbef148-b74e-4d49-ab26-1c1e9808074f" containerName="mariadb-account-create-update" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.971727 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffbef148-b74e-4d49-ab26-1c1e9808074f" containerName="mariadb-account-create-update" Dec 03 08:59:16 crc kubenswrapper[4573]: E1203 08:59:16.971748 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" containerName="dnsmasq-dns" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.971761 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" containerName="dnsmasq-dns" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.972034 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="d62ff626-7df4-4756-ab26-a2d74f71bb57" containerName="mariadb-database-create" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.972079 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" containerName="dnsmasq-dns" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.972096 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffbef148-b74e-4d49-ab26-1c1e9808074f" containerName="mariadb-account-create-update" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.972984 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.985778 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 08:59:16 crc kubenswrapper[4573]: I1203 08:59:16.989286 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gx6qj" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.005438 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8t69f"] Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.142482 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-config-data\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.142614 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-combined-ca-bundle\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.142660 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-db-sync-config-data\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.142722 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c45k\" (UniqueName: \"kubernetes.io/projected/de65a9c9-bef6-42e3-b256-f94d4af37c96-kube-api-access-9c45k\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.247757 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-combined-ca-bundle\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.247869 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-db-sync-config-data\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.247961 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c45k\" (UniqueName: \"kubernetes.io/projected/de65a9c9-bef6-42e3-b256-f94d4af37c96-kube-api-access-9c45k\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.248094 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-config-data\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.254234 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-combined-ca-bundle\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.258829 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-db-sync-config-data\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.261629 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-config-data\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.289502 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c45k\" (UniqueName: \"kubernetes.io/projected/de65a9c9-bef6-42e3-b256-f94d4af37c96-kube-api-access-9c45k\") pod \"glance-db-sync-8t69f\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.298994 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8t69f" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.467293 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-h42zx" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.519034 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-lzbxl" event={"ID":"5ceed462-6691-4229-b764-b25eb42bbf86","Type":"ContainerStarted","Data":"59a396d0d350f29b7ef54d0a82d6d91800f590cd6317cd98f0dd88019ef9d1be"} Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.539844 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-h42zx"] Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.547304 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-h42zx"] Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.594871 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-lzbxl" podStartSLOduration=3.406532029 podStartE2EDuration="9.594813023s" podCreationTimestamp="2025-12-03 08:59:08 +0000 UTC" firstStartedPulling="2025-12-03 08:59:09.964691843 +0000 UTC m=+1270.533071102" lastFinishedPulling="2025-12-03 08:59:16.152972837 +0000 UTC m=+1276.721352096" observedRunningTime="2025-12-03 08:59:17.574102447 +0000 UTC m=+1278.142481726" watchObservedRunningTime="2025-12-03 08:59:17.594813023 +0000 UTC m=+1278.163192292" Dec 03 08:59:17 crc kubenswrapper[4573]: I1203 08:59:17.788602 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 08:59:18 crc kubenswrapper[4573]: I1203 08:59:18.056706 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12f89059-a1f2-45fa-be11-e4bdc6b6f936" path="/var/lib/kubelet/pods/12f89059-a1f2-45fa-be11-e4bdc6b6f936/volumes" Dec 03 08:59:18 crc kubenswrapper[4573]: I1203 08:59:18.057744 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8t69f"] Dec 03 08:59:18 crc kubenswrapper[4573]: W1203 08:59:18.065510 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde65a9c9_bef6_42e3_b256_f94d4af37c96.slice/crio-07b8fd8767b51755d0dfcfb57e94ba1e1a90c6c683f8b62ebb151ec9206f3d61 WatchSource:0}: Error finding container 07b8fd8767b51755d0dfcfb57e94ba1e1a90c6c683f8b62ebb151ec9206f3d61: Status 404 returned error can't find the container with id 07b8fd8767b51755d0dfcfb57e94ba1e1a90c6c683f8b62ebb151ec9206f3d61 Dec 03 08:59:18 crc kubenswrapper[4573]: I1203 08:59:18.531904 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8t69f" event={"ID":"de65a9c9-bef6-42e3-b256-f94d4af37c96","Type":"ContainerStarted","Data":"07b8fd8767b51755d0dfcfb57e94ba1e1a90c6c683f8b62ebb151ec9206f3d61"} Dec 03 08:59:20 crc kubenswrapper[4573]: I1203 08:59:20.552341 4573 generic.go:334] "Generic (PLEG): container finished" podID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerID="b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca" exitCode=0 Dec 03 08:59:20 crc kubenswrapper[4573]: I1203 08:59:20.552420 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9988868b-f38f-4e95-af61-1834e0e67a8b","Type":"ContainerDied","Data":"b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca"} Dec 03 08:59:20 crc kubenswrapper[4573]: I1203 08:59:20.555346 4573 generic.go:334] "Generic (PLEG): container finished" podID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerID="5f27bfda26561875fed220674d65c66eac1bdc140eca3aa513d193416a640c55" exitCode=0 Dec 03 08:59:20 crc kubenswrapper[4573]: I1203 08:59:20.555387 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7087e292-a9fd-4420-bb4a-eb2d50355082","Type":"ContainerDied","Data":"5f27bfda26561875fed220674d65c66eac1bdc140eca3aa513d193416a640c55"} Dec 03 08:59:21 crc kubenswrapper[4573]: I1203 08:59:21.162796 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:21 crc kubenswrapper[4573]: E1203 08:59:21.163128 4573 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 08:59:21 crc kubenswrapper[4573]: E1203 08:59:21.163371 4573 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 08:59:21 crc kubenswrapper[4573]: E1203 08:59:21.163432 4573 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift podName:053ecb51-80b3-4888-9384-8438ba293ae8 nodeName:}" failed. No retries permitted until 2025-12-03 08:59:37.163411737 +0000 UTC m=+1297.731790996 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift") pod "swift-storage-0" (UID: "053ecb51-80b3-4888-9384-8438ba293ae8") : configmap "swift-ring-files" not found Dec 03 08:59:21 crc kubenswrapper[4573]: I1203 08:59:21.595482 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7087e292-a9fd-4420-bb4a-eb2d50355082","Type":"ContainerStarted","Data":"8fc49a0e39d5014a0145f26bd4d0be966ee9fad4deaaf9c035afcb09e6c961b0"} Dec 03 08:59:21 crc kubenswrapper[4573]: I1203 08:59:21.596218 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 08:59:21 crc kubenswrapper[4573]: I1203 08:59:21.611726 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9988868b-f38f-4e95-af61-1834e0e67a8b","Type":"ContainerStarted","Data":"5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64"} Dec 03 08:59:21 crc kubenswrapper[4573]: I1203 08:59:21.612318 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:59:21 crc kubenswrapper[4573]: I1203 08:59:21.644666 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.605668685 podStartE2EDuration="1m5.644637409s" podCreationTimestamp="2025-12-03 08:58:16 +0000 UTC" firstStartedPulling="2025-12-03 08:58:19.00690018 +0000 UTC m=+1219.575279439" lastFinishedPulling="2025-12-03 08:58:46.045868904 +0000 UTC m=+1246.614248163" observedRunningTime="2025-12-03 08:59:21.638326837 +0000 UTC m=+1282.206706096" watchObservedRunningTime="2025-12-03 08:59:21.644637409 +0000 UTC m=+1282.213016668" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.257855 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bj74b" podUID="33a7f303-d65a-4546-bbb5-1e223d48d847" containerName="ovn-controller" probeResult="failure" output=< Dec 03 08:59:27 crc kubenswrapper[4573]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 08:59:27 crc kubenswrapper[4573]: > Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.370392 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.382449 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-jbx79" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.411314 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=43.57645759 podStartE2EDuration="1m10.411285891s" podCreationTimestamp="2025-12-03 08:58:17 +0000 UTC" firstStartedPulling="2025-12-03 08:58:19.285299824 +0000 UTC m=+1219.853679083" lastFinishedPulling="2025-12-03 08:58:46.120128135 +0000 UTC m=+1246.688507384" observedRunningTime="2025-12-03 08:59:21.680127149 +0000 UTC m=+1282.248506438" watchObservedRunningTime="2025-12-03 08:59:27.411285891 +0000 UTC m=+1287.979665150" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.683881 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bj74b-config-sdrkz"] Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.685971 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.697427 4573 generic.go:334] "Generic (PLEG): container finished" podID="5ceed462-6691-4229-b764-b25eb42bbf86" containerID="59a396d0d350f29b7ef54d0a82d6d91800f590cd6317cd98f0dd88019ef9d1be" exitCode=0 Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.698173 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-lzbxl" event={"ID":"5ceed462-6691-4229-b764-b25eb42bbf86","Type":"ContainerDied","Data":"59a396d0d350f29b7ef54d0a82d6d91800f590cd6317cd98f0dd88019ef9d1be"} Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.711448 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.744273 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-scripts\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.744483 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxrpx\" (UniqueName: \"kubernetes.io/projected/b098593e-325a-44ba-8ee2-e935d184abec-kube-api-access-wxrpx\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.744967 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-additional-scripts\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.745054 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-log-ovn\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.745192 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run-ovn\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.745277 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.772469 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bj74b-config-sdrkz"] Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.846339 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxrpx\" (UniqueName: \"kubernetes.io/projected/b098593e-325a-44ba-8ee2-e935d184abec-kube-api-access-wxrpx\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.846451 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-additional-scripts\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.846480 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-log-ovn\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.846522 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run-ovn\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.846551 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.846620 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-scripts\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.849449 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-scripts\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.850468 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-additional-scripts\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.850810 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-log-ovn\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.850865 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run-ovn\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.850909 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:27 crc kubenswrapper[4573]: I1203 08:59:27.900448 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxrpx\" (UniqueName: \"kubernetes.io/projected/b098593e-325a-44ba-8ee2-e935d184abec-kube-api-access-wxrpx\") pod \"ovn-controller-bj74b-config-sdrkz\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:28 crc kubenswrapper[4573]: I1203 08:59:28.008455 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:32 crc kubenswrapper[4573]: I1203 08:59:32.261993 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bj74b" podUID="33a7f303-d65a-4546-bbb5-1e223d48d847" containerName="ovn-controller" probeResult="failure" output=< Dec 03 08:59:32 crc kubenswrapper[4573]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 08:59:32 crc kubenswrapper[4573]: > Dec 03 08:59:36 crc kubenswrapper[4573]: E1203 08:59:36.384378 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 03 08:59:36 crc kubenswrapper[4573]: E1203 08:59:36.385644 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9c45k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-8t69f_openstack(de65a9c9-bef6-42e3-b256-f94d4af37c96): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 08:59:36 crc kubenswrapper[4573]: E1203 08:59:36.386918 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-8t69f" podUID="de65a9c9-bef6-42e3-b256-f94d4af37c96" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.419771 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.551448 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-combined-ca-bundle\") pod \"5ceed462-6691-4229-b764-b25eb42bbf86\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.551661 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-ring-data-devices\") pod \"5ceed462-6691-4229-b764-b25eb42bbf86\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.551753 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj6zb\" (UniqueName: \"kubernetes.io/projected/5ceed462-6691-4229-b764-b25eb42bbf86-kube-api-access-pj6zb\") pod \"5ceed462-6691-4229-b764-b25eb42bbf86\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.551869 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-swiftconf\") pod \"5ceed462-6691-4229-b764-b25eb42bbf86\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.552111 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ceed462-6691-4229-b764-b25eb42bbf86-etc-swift\") pod \"5ceed462-6691-4229-b764-b25eb42bbf86\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.552170 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-dispersionconf\") pod \"5ceed462-6691-4229-b764-b25eb42bbf86\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.552197 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-scripts\") pod \"5ceed462-6691-4229-b764-b25eb42bbf86\" (UID: \"5ceed462-6691-4229-b764-b25eb42bbf86\") " Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.554102 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5ceed462-6691-4229-b764-b25eb42bbf86" (UID: "5ceed462-6691-4229-b764-b25eb42bbf86"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.568846 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ceed462-6691-4229-b764-b25eb42bbf86-kube-api-access-pj6zb" (OuterVolumeSpecName: "kube-api-access-pj6zb") pod "5ceed462-6691-4229-b764-b25eb42bbf86" (UID: "5ceed462-6691-4229-b764-b25eb42bbf86"). InnerVolumeSpecName "kube-api-access-pj6zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.573847 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ceed462-6691-4229-b764-b25eb42bbf86-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5ceed462-6691-4229-b764-b25eb42bbf86" (UID: "5ceed462-6691-4229-b764-b25eb42bbf86"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.583772 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5ceed462-6691-4229-b764-b25eb42bbf86" (UID: "5ceed462-6691-4229-b764-b25eb42bbf86"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.601519 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-scripts" (OuterVolumeSpecName: "scripts") pod "5ceed462-6691-4229-b764-b25eb42bbf86" (UID: "5ceed462-6691-4229-b764-b25eb42bbf86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.643409 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5ceed462-6691-4229-b764-b25eb42bbf86" (UID: "5ceed462-6691-4229-b764-b25eb42bbf86"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.661090 4573 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.661134 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj6zb\" (UniqueName: \"kubernetes.io/projected/5ceed462-6691-4229-b764-b25eb42bbf86-kube-api-access-pj6zb\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.661151 4573 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.661161 4573 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5ceed462-6691-4229-b764-b25eb42bbf86-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.661170 4573 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.661179 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5ceed462-6691-4229-b764-b25eb42bbf86-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.680433 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ceed462-6691-4229-b764-b25eb42bbf86" (UID: "5ceed462-6691-4229-b764-b25eb42bbf86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:59:36 crc kubenswrapper[4573]: W1203 08:59:36.755848 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb098593e_325a_44ba_8ee2_e935d184abec.slice/crio-53ce597fc3a5e89e020160c52d628db7697e552e33326e7b9c876fa8517d23ab WatchSource:0}: Error finding container 53ce597fc3a5e89e020160c52d628db7697e552e33326e7b9c876fa8517d23ab: Status 404 returned error can't find the container with id 53ce597fc3a5e89e020160c52d628db7697e552e33326e7b9c876fa8517d23ab Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.756904 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bj74b-config-sdrkz"] Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.762736 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ceed462-6691-4229-b764-b25eb42bbf86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.785237 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-lzbxl" event={"ID":"5ceed462-6691-4229-b764-b25eb42bbf86","Type":"ContainerDied","Data":"e477364d9e48d69d2b0ea21df0df222d2f40fbab44e1c9af982cc1f56359e858"} Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.785314 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e477364d9e48d69d2b0ea21df0df222d2f40fbab44e1c9af982cc1f56359e858" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.785265 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lzbxl" Dec 03 08:59:36 crc kubenswrapper[4573]: I1203 08:59:36.790540 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b-config-sdrkz" event={"ID":"b098593e-325a-44ba-8ee2-e935d184abec","Type":"ContainerStarted","Data":"53ce597fc3a5e89e020160c52d628db7697e552e33326e7b9c876fa8517d23ab"} Dec 03 08:59:36 crc kubenswrapper[4573]: E1203 08:59:36.792648 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-8t69f" podUID="de65a9c9-bef6-42e3-b256-f94d4af37c96" Dec 03 08:59:37 crc kubenswrapper[4573]: I1203 08:59:37.171906 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:37 crc kubenswrapper[4573]: I1203 08:59:37.180233 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/053ecb51-80b3-4888-9384-8438ba293ae8-etc-swift\") pod \"swift-storage-0\" (UID: \"053ecb51-80b3-4888-9384-8438ba293ae8\") " pod="openstack/swift-storage-0" Dec 03 08:59:37 crc kubenswrapper[4573]: I1203 08:59:37.223355 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-bj74b" Dec 03 08:59:37 crc kubenswrapper[4573]: I1203 08:59:37.223812 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 08:59:37 crc kubenswrapper[4573]: I1203 08:59:37.762418 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 08:59:37 crc kubenswrapper[4573]: I1203 08:59:37.802004 4573 generic.go:334] "Generic (PLEG): container finished" podID="b098593e-325a-44ba-8ee2-e935d184abec" containerID="f38acc7f0c62619582426dcd3c2f17ed18d6c6e8c070e0ddb5862554c5274a5e" exitCode=0 Dec 03 08:59:37 crc kubenswrapper[4573]: I1203 08:59:37.802122 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b-config-sdrkz" event={"ID":"b098593e-325a-44ba-8ee2-e935d184abec","Type":"ContainerDied","Data":"f38acc7f0c62619582426dcd3c2f17ed18d6c6e8c070e0ddb5862554c5274a5e"} Dec 03 08:59:37 crc kubenswrapper[4573]: I1203 08:59:37.806272 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"9d57285ca381b6c2487696f36b4ada091b8438e041ff0db70d9b64cbe3dafcd1"} Dec 03 08:59:38 crc kubenswrapper[4573]: I1203 08:59:38.165353 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 08:59:38 crc kubenswrapper[4573]: I1203 08:59:38.798056 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 08:59:38 crc kubenswrapper[4573]: I1203 08:59:38.992939 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-hfvkh"] Dec 03 08:59:38 crc kubenswrapper[4573]: E1203 08:59:38.993531 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ceed462-6691-4229-b764-b25eb42bbf86" containerName="swift-ring-rebalance" Dec 03 08:59:38 crc kubenswrapper[4573]: I1203 08:59:38.993552 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ceed462-6691-4229-b764-b25eb42bbf86" containerName="swift-ring-rebalance" Dec 03 08:59:38 crc kubenswrapper[4573]: I1203 08:59:38.993773 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ceed462-6691-4229-b764-b25eb42bbf86" containerName="swift-ring-rebalance" Dec 03 08:59:38 crc kubenswrapper[4573]: I1203 08:59:38.994531 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.064981 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hfvkh"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.110531 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-v5rph"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.115494 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.120886 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-operator-scripts\") pod \"cinder-db-create-hfvkh\" (UID: \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\") " pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.121007 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t4m6\" (UniqueName: \"kubernetes.io/projected/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-kube-api-access-2t4m6\") pod \"cinder-db-create-hfvkh\" (UID: \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\") " pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.163752 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b911-account-create-update-54pl8"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.165302 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.175407 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.180630 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v5rph"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.211315 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b911-account-create-update-54pl8"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.229877 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h472w\" (UniqueName: \"kubernetes.io/projected/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-kube-api-access-h472w\") pod \"barbican-db-create-v5rph\" (UID: \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\") " pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.229984 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t4m6\" (UniqueName: \"kubernetes.io/projected/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-kube-api-access-2t4m6\") pod \"cinder-db-create-hfvkh\" (UID: \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\") " pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.230159 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-operator-scripts\") pod \"cinder-db-create-hfvkh\" (UID: \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\") " pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.230186 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-operator-scripts\") pod \"barbican-db-create-v5rph\" (UID: \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\") " pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.231940 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-operator-scripts\") pod \"cinder-db-create-hfvkh\" (UID: \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\") " pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.305457 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t4m6\" (UniqueName: \"kubernetes.io/projected/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-kube-api-access-2t4m6\") pod \"cinder-db-create-hfvkh\" (UID: \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\") " pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.323637 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.332572 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7797a3f-2e04-44d9-a0f9-9958389b9810-operator-scripts\") pod \"barbican-b911-account-create-update-54pl8\" (UID: \"f7797a3f-2e04-44d9-a0f9-9958389b9810\") " pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.332705 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmgks\" (UniqueName: \"kubernetes.io/projected/f7797a3f-2e04-44d9-a0f9-9958389b9810-kube-api-access-jmgks\") pod \"barbican-b911-account-create-update-54pl8\" (UID: \"f7797a3f-2e04-44d9-a0f9-9958389b9810\") " pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.332746 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-operator-scripts\") pod \"barbican-db-create-v5rph\" (UID: \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\") " pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.332809 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h472w\" (UniqueName: \"kubernetes.io/projected/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-kube-api-access-h472w\") pod \"barbican-db-create-v5rph\" (UID: \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\") " pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.333977 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-operator-scripts\") pod \"barbican-db-create-v5rph\" (UID: \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\") " pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.414168 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-019c-account-create-update-kplbv"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.415675 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.427115 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.434496 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h472w\" (UniqueName: \"kubernetes.io/projected/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-kube-api-access-h472w\") pod \"barbican-db-create-v5rph\" (UID: \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\") " pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.436304 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmgks\" (UniqueName: \"kubernetes.io/projected/f7797a3f-2e04-44d9-a0f9-9958389b9810-kube-api-access-jmgks\") pod \"barbican-b911-account-create-update-54pl8\" (UID: \"f7797a3f-2e04-44d9-a0f9-9958389b9810\") " pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.436428 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7797a3f-2e04-44d9-a0f9-9958389b9810-operator-scripts\") pod \"barbican-b911-account-create-update-54pl8\" (UID: \"f7797a3f-2e04-44d9-a0f9-9958389b9810\") " pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.437275 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7797a3f-2e04-44d9-a0f9-9958389b9810-operator-scripts\") pod \"barbican-b911-account-create-update-54pl8\" (UID: \"f7797a3f-2e04-44d9-a0f9-9958389b9810\") " pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.453177 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.476342 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-019c-account-create-update-kplbv"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.540447 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-operator-scripts\") pod \"cinder-019c-account-create-update-kplbv\" (UID: \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\") " pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.540636 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d84jl\" (UniqueName: \"kubernetes.io/projected/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-kube-api-access-d84jl\") pod \"cinder-019c-account-create-update-kplbv\" (UID: \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\") " pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.558671 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmgks\" (UniqueName: \"kubernetes.io/projected/f7797a3f-2e04-44d9-a0f9-9958389b9810-kube-api-access-jmgks\") pod \"barbican-b911-account-create-update-54pl8\" (UID: \"f7797a3f-2e04-44d9-a0f9-9958389b9810\") " pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.644919 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d84jl\" (UniqueName: \"kubernetes.io/projected/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-kube-api-access-d84jl\") pod \"cinder-019c-account-create-update-kplbv\" (UID: \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\") " pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.645001 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-operator-scripts\") pod \"cinder-019c-account-create-update-kplbv\" (UID: \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\") " pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.645952 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-operator-scripts\") pod \"cinder-019c-account-create-update-kplbv\" (UID: \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\") " pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.719706 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d84jl\" (UniqueName: \"kubernetes.io/projected/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-kube-api-access-d84jl\") pod \"cinder-019c-account-create-update-kplbv\" (UID: \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\") " pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.788148 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.824125 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-hklh5"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.825686 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.856236 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-operator-scripts\") pod \"neutron-db-create-hklh5\" (UID: \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\") " pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.856364 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqmc2\" (UniqueName: \"kubernetes.io/projected/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-kube-api-access-xqmc2\") pod \"neutron-db-create-hklh5\" (UID: \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\") " pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.918067 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.920589 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-m8n6d"] Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.922307 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.944685 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5wmq4" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.945532 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.945701 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.945868 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.958210 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-combined-ca-bundle\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.958299 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqmc2\" (UniqueName: \"kubernetes.io/projected/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-kube-api-access-xqmc2\") pod \"neutron-db-create-hklh5\" (UID: \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\") " pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.958329 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-config-data\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.958357 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzjrz\" (UniqueName: \"kubernetes.io/projected/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-kube-api-access-dzjrz\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.958417 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-operator-scripts\") pod \"neutron-db-create-hklh5\" (UID: \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\") " pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.959323 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-operator-scripts\") pod \"neutron-db-create-hklh5\" (UID: \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\") " pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:39 crc kubenswrapper[4573]: I1203 08:59:39.967232 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hklh5"] Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:39.999984 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.031169 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-m8n6d"] Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060229 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run\") pod \"b098593e-325a-44ba-8ee2-e935d184abec\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060342 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxrpx\" (UniqueName: \"kubernetes.io/projected/b098593e-325a-44ba-8ee2-e935d184abec-kube-api-access-wxrpx\") pod \"b098593e-325a-44ba-8ee2-e935d184abec\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060368 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run-ovn\") pod \"b098593e-325a-44ba-8ee2-e935d184abec\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060429 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-scripts\") pod \"b098593e-325a-44ba-8ee2-e935d184abec\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060492 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-log-ovn\") pod \"b098593e-325a-44ba-8ee2-e935d184abec\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060527 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-additional-scripts\") pod \"b098593e-325a-44ba-8ee2-e935d184abec\" (UID: \"b098593e-325a-44ba-8ee2-e935d184abec\") " Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060660 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-combined-ca-bundle\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060715 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-config-data\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.060738 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzjrz\" (UniqueName: \"kubernetes.io/projected/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-kube-api-access-dzjrz\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.061323 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b098593e-325a-44ba-8ee2-e935d184abec" (UID: "b098593e-325a-44ba-8ee2-e935d184abec"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.062509 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-scripts" (OuterVolumeSpecName: "scripts") pod "b098593e-325a-44ba-8ee2-e935d184abec" (UID: "b098593e-325a-44ba-8ee2-e935d184abec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.062547 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b098593e-325a-44ba-8ee2-e935d184abec" (UID: "b098593e-325a-44ba-8ee2-e935d184abec"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.063293 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b098593e-325a-44ba-8ee2-e935d184abec" (UID: "b098593e-325a-44ba-8ee2-e935d184abec"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.074573 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-config-data\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.074690 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run" (OuterVolumeSpecName: "var-run") pod "b098593e-325a-44ba-8ee2-e935d184abec" (UID: "b098593e-325a-44ba-8ee2-e935d184abec"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.078339 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqmc2\" (UniqueName: \"kubernetes.io/projected/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-kube-api-access-xqmc2\") pod \"neutron-db-create-hklh5\" (UID: \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\") " pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.078567 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b098593e-325a-44ba-8ee2-e935d184abec-kube-api-access-wxrpx" (OuterVolumeSpecName: "kube-api-access-wxrpx") pod "b098593e-325a-44ba-8ee2-e935d184abec" (UID: "b098593e-325a-44ba-8ee2-e935d184abec"). InnerVolumeSpecName "kube-api-access-wxrpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.086924 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-combined-ca-bundle\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.093911 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-74c1-account-create-update-2ssqv"] Dec 03 08:59:40 crc kubenswrapper[4573]: E1203 08:59:40.094604 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b098593e-325a-44ba-8ee2-e935d184abec" containerName="ovn-config" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.094627 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b098593e-325a-44ba-8ee2-e935d184abec" containerName="ovn-config" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.094909 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b098593e-325a-44ba-8ee2-e935d184abec" containerName="ovn-config" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.095608 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-74c1-account-create-update-2ssqv"] Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.095719 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.106427 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.110721 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzjrz\" (UniqueName: \"kubernetes.io/projected/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-kube-api-access-dzjrz\") pod \"keystone-db-sync-m8n6d\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.163996 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.164036 4573 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.164068 4573 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b098593e-325a-44ba-8ee2-e935d184abec-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.164083 4573 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.164093 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxrpx\" (UniqueName: \"kubernetes.io/projected/b098593e-325a-44ba-8ee2-e935d184abec-kube-api-access-wxrpx\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.164103 4573 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b098593e-325a-44ba-8ee2-e935d184abec-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.192028 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.273008 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.276910 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw7dk\" (UniqueName: \"kubernetes.io/projected/38e08930-afea-488b-b008-f9d92b8671de-kube-api-access-dw7dk\") pod \"neutron-74c1-account-create-update-2ssqv\" (UID: \"38e08930-afea-488b-b008-f9d92b8671de\") " pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.276988 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38e08930-afea-488b-b008-f9d92b8671de-operator-scripts\") pod \"neutron-74c1-account-create-update-2ssqv\" (UID: \"38e08930-afea-488b-b008-f9d92b8671de\") " pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.403026 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw7dk\" (UniqueName: \"kubernetes.io/projected/38e08930-afea-488b-b008-f9d92b8671de-kube-api-access-dw7dk\") pod \"neutron-74c1-account-create-update-2ssqv\" (UID: \"38e08930-afea-488b-b008-f9d92b8671de\") " pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.403150 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38e08930-afea-488b-b008-f9d92b8671de-operator-scripts\") pod \"neutron-74c1-account-create-update-2ssqv\" (UID: \"38e08930-afea-488b-b008-f9d92b8671de\") " pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.420328 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38e08930-afea-488b-b008-f9d92b8671de-operator-scripts\") pod \"neutron-74c1-account-create-update-2ssqv\" (UID: \"38e08930-afea-488b-b008-f9d92b8671de\") " pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.476259 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw7dk\" (UniqueName: \"kubernetes.io/projected/38e08930-afea-488b-b008-f9d92b8671de-kube-api-access-dw7dk\") pod \"neutron-74c1-account-create-update-2ssqv\" (UID: \"38e08930-afea-488b-b008-f9d92b8671de\") " pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.726055 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.889039 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b-config-sdrkz" event={"ID":"b098593e-325a-44ba-8ee2-e935d184abec","Type":"ContainerDied","Data":"53ce597fc3a5e89e020160c52d628db7697e552e33326e7b9c876fa8517d23ab"} Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.889125 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53ce597fc3a5e89e020160c52d628db7697e552e33326e7b9c876fa8517d23ab" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.889240 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b-config-sdrkz" Dec 03 08:59:40 crc kubenswrapper[4573]: I1203 08:59:40.963765 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-v5rph"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.099766 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b911-account-create-update-54pl8"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.110635 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hfvkh"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.293282 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hklh5"] Dec 03 08:59:41 crc kubenswrapper[4573]: W1203 08:59:41.320018 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62f89ce0_bc6c_4c50_adcf_3cfbc8c5c781.slice/crio-1959984fde80a6b8bb1250d33abe3450b617422f8aca4ff20891f0f1f4511fc9 WatchSource:0}: Error finding container 1959984fde80a6b8bb1250d33abe3450b617422f8aca4ff20891f0f1f4511fc9: Status 404 returned error can't find the container with id 1959984fde80a6b8bb1250d33abe3450b617422f8aca4ff20891f0f1f4511fc9 Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.320675 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bj74b-config-sdrkz"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.348310 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-bj74b-config-sdrkz"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.359389 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-m8n6d"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.410602 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-019c-account-create-update-kplbv"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.519775 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bj74b-config-6qchf"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.527608 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.564327 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bj74b-config-6qchf"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.569523 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.655899 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-additional-scripts\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.655953 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljgfp\" (UniqueName: \"kubernetes.io/projected/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-kube-api-access-ljgfp\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.656018 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-scripts\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.656063 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run-ovn\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.656105 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.656124 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-log-ovn\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.724256 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-74c1-account-create-update-2ssqv"] Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.758638 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-additional-scripts\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.760403 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljgfp\" (UniqueName: \"kubernetes.io/projected/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-kube-api-access-ljgfp\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.760612 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-scripts\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.760734 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run-ovn\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.760889 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.765892 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-log-ovn\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.761575 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run-ovn\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.766210 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-log-ovn\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.761639 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.760313 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-additional-scripts\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.766939 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-scripts\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.820556 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljgfp\" (UniqueName: \"kubernetes.io/projected/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-kube-api-access-ljgfp\") pod \"ovn-controller-bj74b-config-6qchf\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.933767 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m8n6d" event={"ID":"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654","Type":"ContainerStarted","Data":"e324a8a3682c1748ff440a6188a92a4a2ba7fad06b6589f6372e0d41639ed9e6"} Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.949712 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hklh5" event={"ID":"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781","Type":"ContainerStarted","Data":"1959984fde80a6b8bb1250d33abe3450b617422f8aca4ff20891f0f1f4511fc9"} Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.952164 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-019c-account-create-update-kplbv" event={"ID":"bc73654f-5d5d-4b47-ab85-fd28bd21bed2","Type":"ContainerStarted","Data":"43b013d225dc01f1344d9eb3635c2bab194f9b672b3376b7046a1cbceb2bc51f"} Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.955616 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v5rph" event={"ID":"8ff8212d-0f81-4ecd-88ee-cfcc421a1747","Type":"ContainerStarted","Data":"a8c37657b2f5e149a0ffbf74f589cc9661b59ae72e309ebeecfdf88a8046472a"} Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.955871 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v5rph" event={"ID":"8ff8212d-0f81-4ecd-88ee-cfcc421a1747","Type":"ContainerStarted","Data":"340b1a1e28361d49bd0db2b8d3a91b43d9594289309b76b994e274a05aa69509"} Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.958106 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b911-account-create-update-54pl8" event={"ID":"f7797a3f-2e04-44d9-a0f9-9958389b9810","Type":"ContainerStarted","Data":"b1f3a15761e134daca33c05fc478dcee92ebbe4dc9e6ef43cd52f199f3fe2e1e"} Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.960986 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hfvkh" event={"ID":"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e","Type":"ContainerStarted","Data":"1018de56d10637ce91f2fd6f76a572850cd63f4ac84753bca8cc1328b10d4cf7"} Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.961091 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hfvkh" event={"ID":"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e","Type":"ContainerStarted","Data":"48d567ca9f9641994ba5a8cf3c4b977a4273df8487e75b98fba1ac4edf2e35ca"} Dec 03 08:59:41 crc kubenswrapper[4573]: I1203 08:59:41.983639 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-v5rph" podStartSLOduration=2.983605057 podStartE2EDuration="2.983605057s" podCreationTimestamp="2025-12-03 08:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:41.979293419 +0000 UTC m=+1302.547672678" watchObservedRunningTime="2025-12-03 08:59:41.983605057 +0000 UTC m=+1302.551984326" Dec 03 08:59:42 crc kubenswrapper[4573]: I1203 08:59:42.018130 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:42 crc kubenswrapper[4573]: I1203 08:59:42.053740 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b098593e-325a-44ba-8ee2-e935d184abec" path="/var/lib/kubelet/pods/b098593e-325a-44ba-8ee2-e935d184abec/volumes" Dec 03 08:59:42 crc kubenswrapper[4573]: W1203 08:59:42.539186 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38e08930_afea_488b_b008_f9d92b8671de.slice/crio-9bdea949c109c3e834e831c3d7ffd7bd885be08abba9b0b4b87d7835fb9f3812 WatchSource:0}: Error finding container 9bdea949c109c3e834e831c3d7ffd7bd885be08abba9b0b4b87d7835fb9f3812: Status 404 returned error can't find the container with id 9bdea949c109c3e834e831c3d7ffd7bd885be08abba9b0b4b87d7835fb9f3812 Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.003850 4573 generic.go:334] "Generic (PLEG): container finished" podID="f7797a3f-2e04-44d9-a0f9-9958389b9810" containerID="7f083f03abb690a4d24279e62b5db12da2982605d805b23ec683050373dbb0e5" exitCode=0 Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.003932 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b911-account-create-update-54pl8" event={"ID":"f7797a3f-2e04-44d9-a0f9-9958389b9810","Type":"ContainerDied","Data":"7f083f03abb690a4d24279e62b5db12da2982605d805b23ec683050373dbb0e5"} Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.015139 4573 generic.go:334] "Generic (PLEG): container finished" podID="9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e" containerID="1018de56d10637ce91f2fd6f76a572850cd63f4ac84753bca8cc1328b10d4cf7" exitCode=0 Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.015317 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hfvkh" event={"ID":"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e","Type":"ContainerDied","Data":"1018de56d10637ce91f2fd6f76a572850cd63f4ac84753bca8cc1328b10d4cf7"} Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.025633 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hklh5" event={"ID":"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781","Type":"ContainerStarted","Data":"e2f567c239929f4b9f39cffda960b440fefc7418e809061dabfe475e2ab84f84"} Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.046085 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74c1-account-create-update-2ssqv" event={"ID":"38e08930-afea-488b-b008-f9d92b8671de","Type":"ContainerStarted","Data":"9bdea949c109c3e834e831c3d7ffd7bd885be08abba9b0b4b87d7835fb9f3812"} Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.062308 4573 generic.go:334] "Generic (PLEG): container finished" podID="8ff8212d-0f81-4ecd-88ee-cfcc421a1747" containerID="a8c37657b2f5e149a0ffbf74f589cc9661b59ae72e309ebeecfdf88a8046472a" exitCode=0 Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.062384 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v5rph" event={"ID":"8ff8212d-0f81-4ecd-88ee-cfcc421a1747","Type":"ContainerDied","Data":"a8c37657b2f5e149a0ffbf74f589cc9661b59ae72e309ebeecfdf88a8046472a"} Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.139313 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-hklh5" podStartSLOduration=4.139284096 podStartE2EDuration="4.139284096s" podCreationTimestamp="2025-12-03 08:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:43.118889867 +0000 UTC m=+1303.687269126" watchObservedRunningTime="2025-12-03 08:59:43.139284096 +0000 UTC m=+1303.707663355" Dec 03 08:59:43 crc kubenswrapper[4573]: I1203 08:59:43.458798 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bj74b-config-6qchf"] Dec 03 08:59:43 crc kubenswrapper[4573]: W1203 08:59:43.482663 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6d90d5d_d450_4af3_a14e_9a2c73a0b85f.slice/crio-ace46ef4857a3791c45719a3b7ee79fe13555664e3bc37c6a0e71e0ef4fdbb5a WatchSource:0}: Error finding container ace46ef4857a3791c45719a3b7ee79fe13555664e3bc37c6a0e71e0ef4fdbb5a: Status 404 returned error can't find the container with id ace46ef4857a3791c45719a3b7ee79fe13555664e3bc37c6a0e71e0ef4fdbb5a Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.084228 4573 generic.go:334] "Generic (PLEG): container finished" podID="bc73654f-5d5d-4b47-ab85-fd28bd21bed2" containerID="ade849468c4023a969b366cfc153daa2689586665d15505ccda22d0c6012dfcb" exitCode=0 Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.085098 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-019c-account-create-update-kplbv" event={"ID":"bc73654f-5d5d-4b47-ab85-fd28bd21bed2","Type":"ContainerDied","Data":"ade849468c4023a969b366cfc153daa2689586665d15505ccda22d0c6012dfcb"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.096133 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"d5afa3d395ec87a4f3059cc77c6683f05b3230fb4cd56a303f96052854b0b54e"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.096194 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"16c34931aa46eee3032adba8c04fc463cc029cb55752b444acdd38d0bee8ed38"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.096205 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"5f85416df0896511be3056d07aa7e93a86d8aa6c51378b25ed38c3f74e100190"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.096216 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"c22b02093e716babedef2fab780d5b90b89c9e1699a705c0a19a5a4b5b5d7138"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.111780 4573 generic.go:334] "Generic (PLEG): container finished" podID="38e08930-afea-488b-b008-f9d92b8671de" containerID="52472c809a49a3475105ccfba04e5a080eb833b200db80ba885d49dfed5e4281" exitCode=0 Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.111926 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74c1-account-create-update-2ssqv" event={"ID":"38e08930-afea-488b-b008-f9d92b8671de","Type":"ContainerDied","Data":"52472c809a49a3475105ccfba04e5a080eb833b200db80ba885d49dfed5e4281"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.124561 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b-config-6qchf" event={"ID":"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f","Type":"ContainerStarted","Data":"dd14c68ccd8e3b92c3b298436c0b7a83d1beb287c50d190663b4e748a4d61964"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.124622 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b-config-6qchf" event={"ID":"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f","Type":"ContainerStarted","Data":"ace46ef4857a3791c45719a3b7ee79fe13555664e3bc37c6a0e71e0ef4fdbb5a"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.137778 4573 generic.go:334] "Generic (PLEG): container finished" podID="62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781" containerID="e2f567c239929f4b9f39cffda960b440fefc7418e809061dabfe475e2ab84f84" exitCode=0 Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.137911 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hklh5" event={"ID":"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781","Type":"ContainerDied","Data":"e2f567c239929f4b9f39cffda960b440fefc7418e809061dabfe475e2ab84f84"} Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.172091 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-bj74b-config-6qchf" podStartSLOduration=3.172037382 podStartE2EDuration="3.172037382s" podCreationTimestamp="2025-12-03 08:59:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:44.162791599 +0000 UTC m=+1304.731170858" watchObservedRunningTime="2025-12-03 08:59:44.172037382 +0000 UTC m=+1304.740416661" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.836514 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.840990 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.848012 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.970780 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-operator-scripts\") pod \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\" (UID: \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\") " Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.970973 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7797a3f-2e04-44d9-a0f9-9958389b9810-operator-scripts\") pod \"f7797a3f-2e04-44d9-a0f9-9958389b9810\" (UID: \"f7797a3f-2e04-44d9-a0f9-9958389b9810\") " Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.971025 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h472w\" (UniqueName: \"kubernetes.io/projected/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-kube-api-access-h472w\") pod \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\" (UID: \"8ff8212d-0f81-4ecd-88ee-cfcc421a1747\") " Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.971239 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t4m6\" (UniqueName: \"kubernetes.io/projected/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-kube-api-access-2t4m6\") pod \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\" (UID: \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\") " Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.971285 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-operator-scripts\") pod \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\" (UID: \"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e\") " Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.971320 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmgks\" (UniqueName: \"kubernetes.io/projected/f7797a3f-2e04-44d9-a0f9-9958389b9810-kube-api-access-jmgks\") pod \"f7797a3f-2e04-44d9-a0f9-9958389b9810\" (UID: \"f7797a3f-2e04-44d9-a0f9-9958389b9810\") " Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.972372 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e" (UID: "9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.972381 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8ff8212d-0f81-4ecd-88ee-cfcc421a1747" (UID: "8ff8212d-0f81-4ecd-88ee-cfcc421a1747"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.972416 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7797a3f-2e04-44d9-a0f9-9958389b9810-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7797a3f-2e04-44d9-a0f9-9958389b9810" (UID: "f7797a3f-2e04-44d9-a0f9-9958389b9810"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.978888 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-kube-api-access-2t4m6" (OuterVolumeSpecName: "kube-api-access-2t4m6") pod "9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e" (UID: "9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e"). InnerVolumeSpecName "kube-api-access-2t4m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.981136 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7797a3f-2e04-44d9-a0f9-9958389b9810-kube-api-access-jmgks" (OuterVolumeSpecName: "kube-api-access-jmgks") pod "f7797a3f-2e04-44d9-a0f9-9958389b9810" (UID: "f7797a3f-2e04-44d9-a0f9-9958389b9810"). InnerVolumeSpecName "kube-api-access-jmgks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:44 crc kubenswrapper[4573]: I1203 08:59:44.982293 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-kube-api-access-h472w" (OuterVolumeSpecName: "kube-api-access-h472w") pod "8ff8212d-0f81-4ecd-88ee-cfcc421a1747" (UID: "8ff8212d-0f81-4ecd-88ee-cfcc421a1747"). InnerVolumeSpecName "kube-api-access-h472w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.073664 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h472w\" (UniqueName: \"kubernetes.io/projected/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-kube-api-access-h472w\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.073708 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t4m6\" (UniqueName: \"kubernetes.io/projected/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-kube-api-access-2t4m6\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.073719 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.073730 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmgks\" (UniqueName: \"kubernetes.io/projected/f7797a3f-2e04-44d9-a0f9-9958389b9810-kube-api-access-jmgks\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.073739 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ff8212d-0f81-4ecd-88ee-cfcc421a1747-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.073749 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7797a3f-2e04-44d9-a0f9-9958389b9810-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.162603 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-v5rph" event={"ID":"8ff8212d-0f81-4ecd-88ee-cfcc421a1747","Type":"ContainerDied","Data":"340b1a1e28361d49bd0db2b8d3a91b43d9594289309b76b994e274a05aa69509"} Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.162703 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="340b1a1e28361d49bd0db2b8d3a91b43d9594289309b76b994e274a05aa69509" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.163106 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-v5rph" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.169377 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b911-account-create-update-54pl8" event={"ID":"f7797a3f-2e04-44d9-a0f9-9958389b9810","Type":"ContainerDied","Data":"b1f3a15761e134daca33c05fc478dcee92ebbe4dc9e6ef43cd52f199f3fe2e1e"} Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.169389 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b911-account-create-update-54pl8" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.169476 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1f3a15761e134daca33c05fc478dcee92ebbe4dc9e6ef43cd52f199f3fe2e1e" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.172208 4573 generic.go:334] "Generic (PLEG): container finished" podID="a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" containerID="dd14c68ccd8e3b92c3b298436c0b7a83d1beb287c50d190663b4e748a4d61964" exitCode=0 Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.172318 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b-config-6qchf" event={"ID":"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f","Type":"ContainerDied","Data":"dd14c68ccd8e3b92c3b298436c0b7a83d1beb287c50d190663b4e748a4d61964"} Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.177326 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hfvkh" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.178419 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hfvkh" event={"ID":"9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e","Type":"ContainerDied","Data":"48d567ca9f9641994ba5a8cf3c4b977a4273df8487e75b98fba1ac4edf2e35ca"} Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.178505 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48d567ca9f9641994ba5a8cf3c4b977a4273df8487e75b98fba1ac4edf2e35ca" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.917943 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.919395 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.926327 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.993006 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d84jl\" (UniqueName: \"kubernetes.io/projected/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-kube-api-access-d84jl\") pod \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\" (UID: \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\") " Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.993155 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-operator-scripts\") pod \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\" (UID: \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\") " Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.993280 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqmc2\" (UniqueName: \"kubernetes.io/projected/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-kube-api-access-xqmc2\") pod \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\" (UID: \"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781\") " Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.993330 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-operator-scripts\") pod \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\" (UID: \"bc73654f-5d5d-4b47-ab85-fd28bd21bed2\") " Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.993408 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38e08930-afea-488b-b008-f9d92b8671de-operator-scripts\") pod \"38e08930-afea-488b-b008-f9d92b8671de\" (UID: \"38e08930-afea-488b-b008-f9d92b8671de\") " Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.993442 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw7dk\" (UniqueName: \"kubernetes.io/projected/38e08930-afea-488b-b008-f9d92b8671de-kube-api-access-dw7dk\") pod \"38e08930-afea-488b-b008-f9d92b8671de\" (UID: \"38e08930-afea-488b-b008-f9d92b8671de\") " Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.997441 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781" (UID: "62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:45 crc kubenswrapper[4573]: I1203 08:59:45.998179 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38e08930-afea-488b-b008-f9d92b8671de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "38e08930-afea-488b-b008-f9d92b8671de" (UID: "38e08930-afea-488b-b008-f9d92b8671de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.003565 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38e08930-afea-488b-b008-f9d92b8671de-kube-api-access-dw7dk" (OuterVolumeSpecName: "kube-api-access-dw7dk") pod "38e08930-afea-488b-b008-f9d92b8671de" (UID: "38e08930-afea-488b-b008-f9d92b8671de"). InnerVolumeSpecName "kube-api-access-dw7dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.010692 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc73654f-5d5d-4b47-ab85-fd28bd21bed2" (UID: "bc73654f-5d5d-4b47-ab85-fd28bd21bed2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.011860 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-kube-api-access-xqmc2" (OuterVolumeSpecName: "kube-api-access-xqmc2") pod "62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781" (UID: "62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781"). InnerVolumeSpecName "kube-api-access-xqmc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.019626 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-kube-api-access-d84jl" (OuterVolumeSpecName: "kube-api-access-d84jl") pod "bc73654f-5d5d-4b47-ab85-fd28bd21bed2" (UID: "bc73654f-5d5d-4b47-ab85-fd28bd21bed2"). InnerVolumeSpecName "kube-api-access-d84jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.098321 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.098361 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38e08930-afea-488b-b008-f9d92b8671de-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.098373 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw7dk\" (UniqueName: \"kubernetes.io/projected/38e08930-afea-488b-b008-f9d92b8671de-kube-api-access-dw7dk\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.098431 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d84jl\" (UniqueName: \"kubernetes.io/projected/bc73654f-5d5d-4b47-ab85-fd28bd21bed2-kube-api-access-d84jl\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.098443 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.098470 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqmc2\" (UniqueName: \"kubernetes.io/projected/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781-kube-api-access-xqmc2\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.192669 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-019c-account-create-update-kplbv" event={"ID":"bc73654f-5d5d-4b47-ab85-fd28bd21bed2","Type":"ContainerDied","Data":"43b013d225dc01f1344d9eb3635c2bab194f9b672b3376b7046a1cbceb2bc51f"} Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.193843 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43b013d225dc01f1344d9eb3635c2bab194f9b672b3376b7046a1cbceb2bc51f" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.192829 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-019c-account-create-update-kplbv" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.195570 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-74c1-account-create-update-2ssqv" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.195555 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-74c1-account-create-update-2ssqv" event={"ID":"38e08930-afea-488b-b008-f9d92b8671de","Type":"ContainerDied","Data":"9bdea949c109c3e834e831c3d7ffd7bd885be08abba9b0b4b87d7835fb9f3812"} Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.196334 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bdea949c109c3e834e831c3d7ffd7bd885be08abba9b0b4b87d7835fb9f3812" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.199402 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hklh5" event={"ID":"62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781","Type":"ContainerDied","Data":"1959984fde80a6b8bb1250d33abe3450b617422f8aca4ff20891f0f1f4511fc9"} Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.199447 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1959984fde80a6b8bb1250d33abe3450b617422f8aca4ff20891f0f1f4511fc9" Dec 03 08:59:46 crc kubenswrapper[4573]: I1203 08:59:46.199599 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hklh5" Dec 03 08:59:47 crc kubenswrapper[4573]: I1203 08:59:47.215037 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"5bc8b6378f71754059acab61e73af4d9db29539a9b735e00e78bd55bb9f01720"} Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.559213 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.681074 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-scripts\") pod \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.681828 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-log-ovn\") pod \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.681889 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-additional-scripts\") pod \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.681973 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljgfp\" (UniqueName: \"kubernetes.io/projected/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-kube-api-access-ljgfp\") pod \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.681996 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run-ovn\") pod \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.682013 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" (UID: "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.682106 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run\") pod \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\" (UID: \"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f\") " Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.682848 4573 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.682927 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run" (OuterVolumeSpecName: "var-run") pod "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" (UID: "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.682955 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" (UID: "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.682998 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" (UID: "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.683121 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-scripts" (OuterVolumeSpecName: "scripts") pod "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" (UID: "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.686131 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-kube-api-access-ljgfp" (OuterVolumeSpecName: "kube-api-access-ljgfp") pod "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" (UID: "a6d90d5d-d450-4af3-a14e-9a2c73a0b85f"). InnerVolumeSpecName "kube-api-access-ljgfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.785103 4573 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.785147 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljgfp\" (UniqueName: \"kubernetes.io/projected/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-kube-api-access-ljgfp\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.785162 4573 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.785172 4573 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:49 crc kubenswrapper[4573]: I1203 08:59:49.785181 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.254818 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m8n6d" event={"ID":"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654","Type":"ContainerStarted","Data":"676084a6714120ab873f67d0935d1df84f7b21662a8b80eb74d9ce58f07d9d44"} Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.258534 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"b3eedb461f5ae2d7e8ce78d90ae92cb421d4f819762e5135f37c128fac6929e6"} Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.258571 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"4713411ba6cbadfc8c3c04cefacd3f96ad003df1a2482635aba9a736348e61f0"} Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.258581 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"d2cd4f0fb9452b554f2bc073432afe827d09a9d115bb4b1bd66eb8c53cdd560a"} Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.261406 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bj74b-config-6qchf" event={"ID":"a6d90d5d-d450-4af3-a14e-9a2c73a0b85f","Type":"ContainerDied","Data":"ace46ef4857a3791c45719a3b7ee79fe13555664e3bc37c6a0e71e0ef4fdbb5a"} Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.261458 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ace46ef4857a3791c45719a3b7ee79fe13555664e3bc37c6a0e71e0ef4fdbb5a" Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.261514 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bj74b-config-6qchf" Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.277410 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-m8n6d" podStartSLOduration=3.114758654 podStartE2EDuration="11.277381969s" podCreationTimestamp="2025-12-03 08:59:39 +0000 UTC" firstStartedPulling="2025-12-03 08:59:41.507371271 +0000 UTC m=+1302.075750530" lastFinishedPulling="2025-12-03 08:59:49.669994586 +0000 UTC m=+1310.238373845" observedRunningTime="2025-12-03 08:59:50.273283496 +0000 UTC m=+1310.841662755" watchObservedRunningTime="2025-12-03 08:59:50.277381969 +0000 UTC m=+1310.845761228" Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.690726 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bj74b-config-6qchf"] Dec 03 08:59:50 crc kubenswrapper[4573]: I1203 08:59:50.705768 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-bj74b-config-6qchf"] Dec 03 08:59:52 crc kubenswrapper[4573]: I1203 08:59:52.041412 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" path="/var/lib/kubelet/pods/a6d90d5d-d450-4af3-a14e-9a2c73a0b85f/volumes" Dec 03 08:59:54 crc kubenswrapper[4573]: I1203 08:59:54.329863 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"2d6e25eafd5973d30b8a753b53eca0d0d6c91be08bb1c17e56930eda3daebb47"} Dec 03 08:59:54 crc kubenswrapper[4573]: I1203 08:59:54.330810 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"79ac232f91c09fd34d19f77159d2eed8991e70efbb2d1c61d9616520bd76c8b9"} Dec 03 08:59:54 crc kubenswrapper[4573]: I1203 08:59:54.330830 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"23247d08d303970af1880e0e7a3fd1d9f0dd4776003d20fd3751f7d866e2e630"} Dec 03 08:59:54 crc kubenswrapper[4573]: I1203 08:59:54.330843 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"593b68a92824094bae24aa0638668c1ab2edce99c8a69dd91c7c9c58ed086b45"} Dec 03 08:59:54 crc kubenswrapper[4573]: I1203 08:59:54.333830 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8t69f" event={"ID":"de65a9c9-bef6-42e3-b256-f94d4af37c96","Type":"ContainerStarted","Data":"84dcfceb254e96342219e95412e54d29973b82e2655f344fb62ea90137bce299"} Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.353551 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"50f5344dc3b1e2bdfb3f28b117ef9e49697dd8482fd7c27fd44e16cd4862b8b6"} Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.353619 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"8dc600430181c67ab3a70c8d745cebef8c852c197ee55710b72633ab64369e7f"} Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.353629 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"053ecb51-80b3-4888-9384-8438ba293ae8","Type":"ContainerStarted","Data":"ff724da1c094e2fc1e0624e70cb15cdd3c8e685dea9051cb54bc8fd08b56af70"} Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.404433 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-8t69f" podStartSLOduration=4.388486182 podStartE2EDuration="39.404400957s" podCreationTimestamp="2025-12-03 08:59:16 +0000 UTC" firstStartedPulling="2025-12-03 08:59:18.06855086 +0000 UTC m=+1278.636930119" lastFinishedPulling="2025-12-03 08:59:53.084465645 +0000 UTC m=+1313.652844894" observedRunningTime="2025-12-03 08:59:54.361573515 +0000 UTC m=+1314.929952764" watchObservedRunningTime="2025-12-03 08:59:55.404400957 +0000 UTC m=+1315.972780216" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.406415 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.101800358 podStartE2EDuration="51.406399562s" podCreationTimestamp="2025-12-03 08:59:04 +0000 UTC" firstStartedPulling="2025-12-03 08:59:37.778935255 +0000 UTC m=+1298.347314514" lastFinishedPulling="2025-12-03 08:59:53.083534459 +0000 UTC m=+1313.651913718" observedRunningTime="2025-12-03 08:59:55.396063399 +0000 UTC m=+1315.964442658" watchObservedRunningTime="2025-12-03 08:59:55.406399562 +0000 UTC m=+1315.974778821" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.730508 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lvcdn"] Dec 03 08:59:55 crc kubenswrapper[4573]: E1203 08:59:55.735639 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc73654f-5d5d-4b47-ab85-fd28bd21bed2" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.735719 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc73654f-5d5d-4b47-ab85-fd28bd21bed2" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: E1203 08:59:55.735824 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ff8212d-0f81-4ecd-88ee-cfcc421a1747" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.735831 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ff8212d-0f81-4ecd-88ee-cfcc421a1747" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: E1203 08:59:55.735854 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38e08930-afea-488b-b008-f9d92b8671de" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.735865 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="38e08930-afea-488b-b008-f9d92b8671de" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: E1203 08:59:55.735904 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" containerName="ovn-config" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.735911 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" containerName="ovn-config" Dec 03 08:59:55 crc kubenswrapper[4573]: E1203 08:59:55.735931 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.735938 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: E1203 08:59:55.735956 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7797a3f-2e04-44d9-a0f9-9958389b9810" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.735964 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7797a3f-2e04-44d9-a0f9-9958389b9810" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: E1203 08:59:55.735987 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.735998 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.736804 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ff8212d-0f81-4ecd-88ee-cfcc421a1747" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.736848 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7797a3f-2e04-44d9-a0f9-9958389b9810" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.736873 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="38e08930-afea-488b-b008-f9d92b8671de" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.736893 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.736914 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc73654f-5d5d-4b47-ab85-fd28bd21bed2" containerName="mariadb-account-create-update" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.736938 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d90d5d-d450-4af3-a14e-9a2c73a0b85f" containerName="ovn-config" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.736963 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e" containerName="mariadb-database-create" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.747219 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.751124 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.802842 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lvcdn"] Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.847473 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.847565 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-config\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.847594 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.847706 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.847760 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.847860 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97xxb\" (UniqueName: \"kubernetes.io/projected/06a5e660-c639-4897-8968-d7fabbba27f8-kube-api-access-97xxb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.950040 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.950133 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-config\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.950153 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.950201 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.950234 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.950290 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97xxb\" (UniqueName: \"kubernetes.io/projected/06a5e660-c639-4897-8968-d7fabbba27f8-kube-api-access-97xxb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.951682 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.952272 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-config\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.952886 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.953357 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.953783 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:55 crc kubenswrapper[4573]: I1203 08:59:55.984499 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97xxb\" (UniqueName: \"kubernetes.io/projected/06a5e660-c639-4897-8968-d7fabbba27f8-kube-api-access-97xxb\") pod \"dnsmasq-dns-5c79d794d7-lvcdn\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:56 crc kubenswrapper[4573]: I1203 08:59:56.083029 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:56 crc kubenswrapper[4573]: I1203 08:59:56.368449 4573 generic.go:334] "Generic (PLEG): container finished" podID="09ca3bd5-ac7a-40c1-b9b9-394ee79f5654" containerID="676084a6714120ab873f67d0935d1df84f7b21662a8b80eb74d9ce58f07d9d44" exitCode=0 Dec 03 08:59:56 crc kubenswrapper[4573]: I1203 08:59:56.369607 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m8n6d" event={"ID":"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654","Type":"ContainerDied","Data":"676084a6714120ab873f67d0935d1df84f7b21662a8b80eb74d9ce58f07d9d44"} Dec 03 08:59:56 crc kubenswrapper[4573]: I1203 08:59:56.622164 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lvcdn"] Dec 03 08:59:57 crc kubenswrapper[4573]: I1203 08:59:57.404189 4573 generic.go:334] "Generic (PLEG): container finished" podID="06a5e660-c639-4897-8968-d7fabbba27f8" containerID="c82fa8904efaaedcee47ae73fa964bc7a84aa459165085b8eb2d1feec6a16948" exitCode=0 Dec 03 08:59:57 crc kubenswrapper[4573]: I1203 08:59:57.405786 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" event={"ID":"06a5e660-c639-4897-8968-d7fabbba27f8","Type":"ContainerDied","Data":"c82fa8904efaaedcee47ae73fa964bc7a84aa459165085b8eb2d1feec6a16948"} Dec 03 08:59:57 crc kubenswrapper[4573]: I1203 08:59:57.405829 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" event={"ID":"06a5e660-c639-4897-8968-d7fabbba27f8","Type":"ContainerStarted","Data":"20393ea09a315a8ea7d23c7534114c3b5f60f37523c7b4bb3a6762018d9739ee"} Dec 03 08:59:57 crc kubenswrapper[4573]: I1203 08:59:57.916936 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.115844 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-config-data\") pod \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.116093 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzjrz\" (UniqueName: \"kubernetes.io/projected/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-kube-api-access-dzjrz\") pod \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.116220 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-combined-ca-bundle\") pod \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\" (UID: \"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654\") " Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.151386 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-kube-api-access-dzjrz" (OuterVolumeSpecName: "kube-api-access-dzjrz") pod "09ca3bd5-ac7a-40c1-b9b9-394ee79f5654" (UID: "09ca3bd5-ac7a-40c1-b9b9-394ee79f5654"). InnerVolumeSpecName "kube-api-access-dzjrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.219289 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09ca3bd5-ac7a-40c1-b9b9-394ee79f5654" (UID: "09ca3bd5-ac7a-40c1-b9b9-394ee79f5654"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.243817 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzjrz\" (UniqueName: \"kubernetes.io/projected/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-kube-api-access-dzjrz\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.243868 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.310389 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-config-data" (OuterVolumeSpecName: "config-data") pod "09ca3bd5-ac7a-40c1-b9b9-394ee79f5654" (UID: "09ca3bd5-ac7a-40c1-b9b9-394ee79f5654"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.345998 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.418193 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m8n6d" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.418539 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m8n6d" event={"ID":"09ca3bd5-ac7a-40c1-b9b9-394ee79f5654","Type":"ContainerDied","Data":"e324a8a3682c1748ff440a6188a92a4a2ba7fad06b6589f6372e0d41639ed9e6"} Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.418632 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e324a8a3682c1748ff440a6188a92a4a2ba7fad06b6589f6372e0d41639ed9e6" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.421075 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" event={"ID":"06a5e660-c639-4897-8968-d7fabbba27f8","Type":"ContainerStarted","Data":"a6ce83d0bf0f20f9f5180e3e8c004ceb82d9c0804bde642e97b8c68721816c03"} Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.421301 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.453693 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" podStartSLOduration=3.4536562760000002 podStartE2EDuration="3.453656276s" podCreationTimestamp="2025-12-03 08:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 08:59:58.445901524 +0000 UTC m=+1319.014280783" watchObservedRunningTime="2025-12-03 08:59:58.453656276 +0000 UTC m=+1319.022035535" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.727085 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lvcdn"] Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.751213 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v556f"] Dec 03 08:59:58 crc kubenswrapper[4573]: E1203 08:59:58.751679 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ca3bd5-ac7a-40c1-b9b9-394ee79f5654" containerName="keystone-db-sync" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.751723 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ca3bd5-ac7a-40c1-b9b9-394ee79f5654" containerName="keystone-db-sync" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.751949 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ca3bd5-ac7a-40c1-b9b9-394ee79f5654" containerName="keystone-db-sync" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.753193 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.770209 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v556f"] Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.770938 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.794715 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5wmq4" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.795137 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.795268 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.809445 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.848546 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-zwp5l"] Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.850685 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.876591 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56tsq\" (UniqueName: \"kubernetes.io/projected/a051e848-8cfe-4ea3-bed8-721665a5a37b-kube-api-access-56tsq\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.876880 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-scripts\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.876954 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-credential-keys\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.876989 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.877222 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-config\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.877258 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-config-data\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.877379 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-svc\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.877502 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-fernet-keys\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.877558 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.877625 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-combined-ca-bundle\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.886315 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.886546 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ggpn\" (UniqueName: \"kubernetes.io/projected/7f764a7a-b747-4b12-b91d-093ee963d8cb-kube-api-access-7ggpn\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:58 crc kubenswrapper[4573]: I1203 08:59:58.942270 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-zwp5l"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998405 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-svc\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998494 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-fernet-keys\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998518 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998537 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-combined-ca-bundle\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998592 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998620 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ggpn\" (UniqueName: \"kubernetes.io/projected/7f764a7a-b747-4b12-b91d-093ee963d8cb-kube-api-access-7ggpn\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998658 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56tsq\" (UniqueName: \"kubernetes.io/projected/a051e848-8cfe-4ea3-bed8-721665a5a37b-kube-api-access-56tsq\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998700 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-scripts\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998718 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-credential-keys\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998735 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998771 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-config\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:58.998788 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-config-data\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.000664 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-svc\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.010177 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.011447 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.011457 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-config\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.012387 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.012739 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-config-data\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.014857 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-scripts\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.044635 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-fernet-keys\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.045184 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-credential-keys\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.054893 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-combined-ca-bundle\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.095970 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56tsq\" (UniqueName: \"kubernetes.io/projected/a051e848-8cfe-4ea3-bed8-721665a5a37b-kube-api-access-56tsq\") pod \"keystone-bootstrap-v556f\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.102996 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ggpn\" (UniqueName: \"kubernetes.io/projected/7f764a7a-b747-4b12-b91d-093ee963d8cb-kube-api-access-7ggpn\") pod \"dnsmasq-dns-5b868669f-zwp5l\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.183755 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.342535 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-chs2b"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.357601 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.378121 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v556f" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.403211 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-qmt4b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.403506 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.403643 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.412497 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwfgd\" (UniqueName: \"kubernetes.io/projected/5c73103b-e4fd-4a54-ad2c-021db8042b13-kube-api-access-vwfgd\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.412565 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-db-sync-config-data\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.412607 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c73103b-e4fd-4a54-ad2c-021db8042b13-etc-machine-id\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.412652 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-combined-ca-bundle\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.412706 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-config-data\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.412758 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-scripts\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.508019 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-chs2b"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.517752 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c73103b-e4fd-4a54-ad2c-021db8042b13-etc-machine-id\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.517854 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-combined-ca-bundle\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.517914 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-config-data\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.517959 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-scripts\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.518021 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwfgd\" (UniqueName: \"kubernetes.io/projected/5c73103b-e4fd-4a54-ad2c-021db8042b13-kube-api-access-vwfgd\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.518069 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-db-sync-config-data\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.520237 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c73103b-e4fd-4a54-ad2c-021db8042b13-etc-machine-id\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.534040 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-db-sync-config-data\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.537409 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-config-data\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.559652 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6c795868d9-6tqff"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.559996 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-combined-ca-bundle\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.572135 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.581269 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-scripts\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.612305 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c795868d9-6tqff"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.621271 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5451c2e0-1168-4793-a1ba-c07db67f7b8a-horizon-secret-key\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.621356 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-scripts\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.622005 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kkcl\" (UniqueName: \"kubernetes.io/projected/5451c2e0-1168-4793-a1ba-c07db67f7b8a-kube-api-access-7kkcl\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.622065 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-config-data\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.622113 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5451c2e0-1168-4793-a1ba-c07db67f7b8a-logs\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.630088 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwfgd\" (UniqueName: \"kubernetes.io/projected/5c73103b-e4fd-4a54-ad2c-021db8042b13-kube-api-access-vwfgd\") pod \"cinder-db-sync-chs2b\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.631599 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.631869 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.632293 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-k9gnt" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.632653 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.694129 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.696869 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.712997 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.713903 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.724414 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-srtdq"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.725968 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-srtdq" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728365 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728407 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5451c2e0-1168-4793-a1ba-c07db67f7b8a-logs\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728445 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-scripts\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728474 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5451c2e0-1168-4793-a1ba-c07db67f7b8a-horizon-secret-key\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728514 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-scripts\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728534 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb5sz\" (UniqueName: \"kubernetes.io/projected/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-kube-api-access-rb5sz\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728580 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-config-data\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728613 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728640 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-log-httpd\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728660 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-run-httpd\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728679 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kkcl\" (UniqueName: \"kubernetes.io/projected/5451c2e0-1168-4793-a1ba-c07db67f7b8a-kube-api-access-7kkcl\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.728707 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-config-data\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.741140 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-config-data\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.744671 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5451c2e0-1168-4793-a1ba-c07db67f7b8a-logs\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.746384 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-scripts\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.751675 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.751963 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.752098 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-255sb" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.766989 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-fnppb"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.769874 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-chs2b" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.782952 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5451c2e0-1168-4793-a1ba-c07db67f7b8a-horizon-secret-key\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.783466 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fnppb" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.794885 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kkcl\" (UniqueName: \"kubernetes.io/projected/5451c2e0-1168-4793-a1ba-c07db67f7b8a-kube-api-access-7kkcl\") pod \"horizon-6c795868d9-6tqff\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " pod="openstack/horizon-6c795868d9-6tqff" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.796323 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.808377 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jkqdd" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.811262 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-fnppb"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834381 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb5sz\" (UniqueName: \"kubernetes.io/projected/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-kube-api-access-rb5sz\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834461 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-combined-ca-bundle\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834499 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-combined-ca-bundle\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834530 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxl4r\" (UniqueName: \"kubernetes.io/projected/3cd84894-b4fa-481f-8856-678218f7bad7-kube-api-access-xxl4r\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834588 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-db-sync-config-data\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834622 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-config-data\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834707 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-config\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834770 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834795 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nxhw\" (UniqueName: \"kubernetes.io/projected/8e8386e5-315f-4d6f-a560-e1a45777e339-kube-api-access-5nxhw\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834826 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-log-httpd\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834858 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-run-httpd\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834905 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.834963 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-scripts\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.836568 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-log-httpd\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.837700 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-run-httpd\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.864713 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-scripts\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.864902 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-config-data\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.872876 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.887410 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.890226 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-srtdq"] Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.951359 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-combined-ca-bundle\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.951435 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxl4r\" (UniqueName: \"kubernetes.io/projected/3cd84894-b4fa-481f-8856-678218f7bad7-kube-api-access-xxl4r\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.951462 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-combined-ca-bundle\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.951533 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-db-sync-config-data\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.951593 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-config\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.951645 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nxhw\" (UniqueName: \"kubernetes.io/projected/8e8386e5-315f-4d6f-a560-e1a45777e339-kube-api-access-5nxhw\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.979651 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.986680 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb5sz\" (UniqueName: \"kubernetes.io/projected/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-kube-api-access-rb5sz\") pod \"ceilometer-0\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " pod="openstack/ceilometer-0" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.987214 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-k9gnt" Dec 03 08:59:59 crc kubenswrapper[4573]: I1203 08:59:59.993767 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c795868d9-6tqff" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.001978 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-combined-ca-bundle\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.002074 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.002884 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.028146 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-combined-ca-bundle\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.029704 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-db-sync-config-data\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.050342 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nxhw\" (UniqueName: \"kubernetes.io/projected/8e8386e5-315f-4d6f-a560-e1a45777e339-kube-api-access-5nxhw\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.071771 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-config\") pod \"neutron-db-sync-srtdq\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " pod="openstack/neutron-db-sync-srtdq" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.105072 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.123222 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-255sb" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.127275 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-srtdq" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.151412 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxl4r\" (UniqueName: \"kubernetes.io/projected/3cd84894-b4fa-481f-8856-678218f7bad7-kube-api-access-xxl4r\") pod \"barbican-db-sync-fnppb\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " pod="openstack/barbican-db-sync-fnppb" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.177134 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jkqdd" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.194720 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fnppb" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.274245 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-zwp5l"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.424117 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-jckvv"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.446172 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.623768 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" podUID="06a5e660-c639-4897-8968-d7fabbba27f8" containerName="dnsmasq-dns" containerID="cri-o://a6ce83d0bf0f20f9f5180e3e8c004ceb82d9c0804bde642e97b8c68721816c03" gracePeriod=10 Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.637698 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-kvbg6"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.644340 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.656638 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-svc\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.656792 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-config\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.656844 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9krcj\" (UniqueName: \"kubernetes.io/projected/1f1b04c5-a752-4557-a537-ad477d6d6280-kube-api-access-9krcj\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.656947 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.657081 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.657186 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.661185 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.661564 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vtbrw" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.661443 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.665623 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-jckvv"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767113 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-config\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767171 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9krcj\" (UniqueName: \"kubernetes.io/projected/1f1b04c5-a752-4557-a537-ad477d6d6280-kube-api-access-9krcj\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767214 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767258 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hhsj\" (UniqueName: \"kubernetes.io/projected/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-kube-api-access-4hhsj\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767289 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767319 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-combined-ca-bundle\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767346 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-config-data\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767364 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767408 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-scripts\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767435 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-svc\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.767466 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-logs\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.768541 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-config\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.770631 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.771374 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.771966 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-svc\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.779189 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.801080 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d7849547c-j77g9"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.816254 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.830958 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9krcj\" (UniqueName: \"kubernetes.io/projected/1f1b04c5-a752-4557-a537-ad477d6d6280-kube-api-access-9krcj\") pod \"dnsmasq-dns-cf78879c9-jckvv\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.832607 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-kvbg6"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.875348 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d7849547c-j77g9"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.876793 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-scripts\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.876872 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-logs\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.876916 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-scripts\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.876952 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e511768a-604b-441d-a363-e5c522b39647-horizon-secret-key\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.876985 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e511768a-604b-441d-a363-e5c522b39647-logs\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.877024 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84ddg\" (UniqueName: \"kubernetes.io/projected/e511768a-604b-441d-a363-e5c522b39647-kube-api-access-84ddg\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.877076 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-config-data\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.877135 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hhsj\" (UniqueName: \"kubernetes.io/projected/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-kube-api-access-4hhsj\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.877185 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-combined-ca-bundle\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.877220 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-config-data\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.888245 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-logs\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.909220 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-combined-ca-bundle\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.909310 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv"] Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.911118 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:00 crc kubenswrapper[4573]: I1203 09:00:00.918043 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-config-data\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.934702 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.935123 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.935969 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-scripts\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.952181 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv"] Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.993753 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-scripts\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.993805 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e511768a-604b-441d-a363-e5c522b39647-horizon-secret-key\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.993833 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e511768a-604b-441d-a363-e5c522b39647-logs\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.993865 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/561700fb-a6d4-42e1-b339-f3dd062a504a-config-volume\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.993897 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84ddg\" (UniqueName: \"kubernetes.io/projected/e511768a-604b-441d-a363-e5c522b39647-kube-api-access-84ddg\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.993924 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-config-data\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.993951 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp8kw\" (UniqueName: \"kubernetes.io/projected/561700fb-a6d4-42e1-b339-f3dd062a504a-kube-api-access-fp8kw\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.993980 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/561700fb-a6d4-42e1-b339-f3dd062a504a-secret-volume\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.994877 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-scripts\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.995968 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e511768a-604b-441d-a363-e5c522b39647-logs\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:00.996995 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-config-data\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.010299 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.011656 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hhsj\" (UniqueName: \"kubernetes.io/projected/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-kube-api-access-4hhsj\") pod \"placement-db-sync-kvbg6\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.051169 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-zwp5l"] Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.103497 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/561700fb-a6d4-42e1-b339-f3dd062a504a-config-volume\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.103571 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/561700fb-a6d4-42e1-b339-f3dd062a504a-secret-volume\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.103592 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp8kw\" (UniqueName: \"kubernetes.io/projected/561700fb-a6d4-42e1-b339-f3dd062a504a-kube-api-access-fp8kw\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.104921 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/561700fb-a6d4-42e1-b339-f3dd062a504a-config-volume\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.107457 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-kvbg6" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.120286 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84ddg\" (UniqueName: \"kubernetes.io/projected/e511768a-604b-441d-a363-e5c522b39647-kube-api-access-84ddg\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.120665 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e511768a-604b-441d-a363-e5c522b39647-horizon-secret-key\") pod \"horizon-7d7849547c-j77g9\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.145446 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/561700fb-a6d4-42e1-b339-f3dd062a504a-secret-volume\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.226517 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.240927 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp8kw\" (UniqueName: \"kubernetes.io/projected/561700fb-a6d4-42e1-b339-f3dd062a504a-kube-api-access-fp8kw\") pod \"collect-profiles-29412540-2ddxv\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.368496 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v556f"] Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.489857 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.555480 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.754298 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96" containerName="galera" probeResult="failure" output="command timed out" Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.767657 4573 generic.go:334] "Generic (PLEG): container finished" podID="06a5e660-c639-4897-8968-d7fabbba27f8" containerID="a6ce83d0bf0f20f9f5180e3e8c004ceb82d9c0804bde642e97b8c68721816c03" exitCode=0 Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.767779 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" event={"ID":"06a5e660-c639-4897-8968-d7fabbba27f8","Type":"ContainerDied","Data":"a6ce83d0bf0f20f9f5180e3e8c004ceb82d9c0804bde642e97b8c68721816c03"} Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.774399 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v556f" event={"ID":"a051e848-8cfe-4ea3-bed8-721665a5a37b","Type":"ContainerStarted","Data":"ddb53d54270fd9eb922defe011498666c77a7d1dbaba4ce470434f5f44bd27f7"} Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:01.806121 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-zwp5l" event={"ID":"7f764a7a-b747-4b12-b91d-093ee963d8cb","Type":"ContainerStarted","Data":"74a15500ed350db7d03665de3338f8e3972b763c0d25fe46e6ca8a2b21adfb62"} Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:02.821622 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v556f" event={"ID":"a051e848-8cfe-4ea3-bed8-721665a5a37b","Type":"ContainerStarted","Data":"c5991229e5c00013cea09dc1f612403384de8e841bcf82e821749bd6642ea1a4"} Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:02.828430 4573 generic.go:334] "Generic (PLEG): container finished" podID="7f764a7a-b747-4b12-b91d-093ee963d8cb" containerID="f89ebb241acb65b1941be33886abf87153532abd430f9e49e64e2b85561497e6" exitCode=0 Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:02.828505 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-zwp5l" event={"ID":"7f764a7a-b747-4b12-b91d-093ee963d8cb","Type":"ContainerDied","Data":"f89ebb241acb65b1941be33886abf87153532abd430f9e49e64e2b85561497e6"} Dec 03 09:00:02 crc kubenswrapper[4573]: I1203 09:00:02.866763 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v556f" podStartSLOduration=4.866739728 podStartE2EDuration="4.866739728s" podCreationTimestamp="2025-12-03 08:59:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:00:02.853016323 +0000 UTC m=+1323.421395582" watchObservedRunningTime="2025-12-03 09:00:02.866739728 +0000 UTC m=+1323.435118987" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.502742 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-chs2b"] Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.589693 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c795868d9-6tqff"] Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.605160 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.628846 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bdd77d6bf-c2wp4"] Dec 03 09:00:03 crc kubenswrapper[4573]: E1203 09:00:03.629382 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06a5e660-c639-4897-8968-d7fabbba27f8" containerName="dnsmasq-dns" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.629405 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="06a5e660-c639-4897-8968-d7fabbba27f8" containerName="dnsmasq-dns" Dec 03 09:00:03 crc kubenswrapper[4573]: E1203 09:00:03.629422 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06a5e660-c639-4897-8968-d7fabbba27f8" containerName="init" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.629429 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="06a5e660-c639-4897-8968-d7fabbba27f8" containerName="init" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.629659 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="06a5e660-c639-4897-8968-d7fabbba27f8" containerName="dnsmasq-dns" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.630858 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.670182 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bdd77d6bf-c2wp4"] Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.687655 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-swift-storage-0\") pod \"06a5e660-c639-4897-8968-d7fabbba27f8\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.687752 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-svc\") pod \"06a5e660-c639-4897-8968-d7fabbba27f8\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.687831 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-sb\") pod \"06a5e660-c639-4897-8968-d7fabbba27f8\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.687909 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-nb\") pod \"06a5e660-c639-4897-8968-d7fabbba27f8\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.687971 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97xxb\" (UniqueName: \"kubernetes.io/projected/06a5e660-c639-4897-8968-d7fabbba27f8-kube-api-access-97xxb\") pod \"06a5e660-c639-4897-8968-d7fabbba27f8\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.688025 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-config\") pod \"06a5e660-c639-4897-8968-d7fabbba27f8\" (UID: \"06a5e660-c639-4897-8968-d7fabbba27f8\") " Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.688188 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7f6n\" (UniqueName: \"kubernetes.io/projected/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-kube-api-access-c7f6n\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.688263 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-logs\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.688312 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-horizon-secret-key\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.688360 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-config-data\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.688389 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-scripts\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.775783 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06a5e660-c639-4897-8968-d7fabbba27f8-kube-api-access-97xxb" (OuterVolumeSpecName: "kube-api-access-97xxb") pod "06a5e660-c639-4897-8968-d7fabbba27f8" (UID: "06a5e660-c639-4897-8968-d7fabbba27f8"). InnerVolumeSpecName "kube-api-access-97xxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.797234 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7f6n\" (UniqueName: \"kubernetes.io/projected/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-kube-api-access-c7f6n\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.797342 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-logs\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.797396 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-horizon-secret-key\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.797444 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-config-data\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.797472 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-scripts\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.800676 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97xxb\" (UniqueName: \"kubernetes.io/projected/06a5e660-c639-4897-8968-d7fabbba27f8-kube-api-access-97xxb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.802006 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-logs\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.804026 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-scripts\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.817369 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-config-data\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.854390 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" event={"ID":"06a5e660-c639-4897-8968-d7fabbba27f8","Type":"ContainerDied","Data":"20393ea09a315a8ea7d23c7534114c3b5f60f37523c7b4bb3a6762018d9739ee"} Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.854453 4573 scope.go:117] "RemoveContainer" containerID="a6ce83d0bf0f20f9f5180e3e8c004ceb82d9c0804bde642e97b8c68721816c03" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.854581 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-lvcdn" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.863159 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-horizon-secret-key\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.918949 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-chs2b" event={"ID":"5c73103b-e4fd-4a54-ad2c-021db8042b13","Type":"ContainerStarted","Data":"01feccc9862109d0a879828bf169aa6fb844d8f0240db330650b322ea3407d7f"} Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.948768 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7f6n\" (UniqueName: \"kubernetes.io/projected/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-kube-api-access-c7f6n\") pod \"horizon-5bdd77d6bf-c2wp4\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:03 crc kubenswrapper[4573]: I1203 09:00:03.971291 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "06a5e660-c639-4897-8968-d7fabbba27f8" (UID: "06a5e660-c639-4897-8968-d7fabbba27f8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.002786 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-config" (OuterVolumeSpecName: "config") pod "06a5e660-c639-4897-8968-d7fabbba27f8" (UID: "06a5e660-c639-4897-8968-d7fabbba27f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.008290 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.028514 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.028709 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.028998 4573 scope.go:117] "RemoveContainer" containerID="c82fa8904efaaedcee47ae73fa964bc7a84aa459165085b8eb2d1feec6a16948" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.031673 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.104405 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "06a5e660-c639-4897-8968-d7fabbba27f8" (UID: "06a5e660-c639-4897-8968-d7fabbba27f8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.109510 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c795868d9-6tqff"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.109532 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "06a5e660-c639-4897-8968-d7fabbba27f8" (UID: "06a5e660-c639-4897-8968-d7fabbba27f8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.115841 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "06a5e660-c639-4897-8968-d7fabbba27f8" (UID: "06a5e660-c639-4897-8968-d7fabbba27f8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.184797 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.185440 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.185547 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06a5e660-c639-4897-8968-d7fabbba27f8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.237598 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.257797 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-fnppb"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.294184 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-nb\") pod \"7f764a7a-b747-4b12-b91d-093ee963d8cb\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.294299 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-svc\") pod \"7f764a7a-b747-4b12-b91d-093ee963d8cb\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.294380 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-sb\") pod \"7f764a7a-b747-4b12-b91d-093ee963d8cb\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.294424 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-swift-storage-0\") pod \"7f764a7a-b747-4b12-b91d-093ee963d8cb\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.294529 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ggpn\" (UniqueName: \"kubernetes.io/projected/7f764a7a-b747-4b12-b91d-093ee963d8cb-kube-api-access-7ggpn\") pod \"7f764a7a-b747-4b12-b91d-093ee963d8cb\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.294793 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-config\") pod \"7f764a7a-b747-4b12-b91d-093ee963d8cb\" (UID: \"7f764a7a-b747-4b12-b91d-093ee963d8cb\") " Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.375727 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f764a7a-b747-4b12-b91d-093ee963d8cb-kube-api-access-7ggpn" (OuterVolumeSpecName: "kube-api-access-7ggpn") pod "7f764a7a-b747-4b12-b91d-093ee963d8cb" (UID: "7f764a7a-b747-4b12-b91d-093ee963d8cb"). InnerVolumeSpecName "kube-api-access-7ggpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.387741 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-config" (OuterVolumeSpecName: "config") pod "7f764a7a-b747-4b12-b91d-093ee963d8cb" (UID: "7f764a7a-b747-4b12-b91d-093ee963d8cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.403029 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ggpn\" (UniqueName: \"kubernetes.io/projected/7f764a7a-b747-4b12-b91d-093ee963d8cb-kube-api-access-7ggpn\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.403205 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.403634 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-srtdq"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.544418 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-kvbg6"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.589628 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f764a7a-b747-4b12-b91d-093ee963d8cb" (UID: "7f764a7a-b747-4b12-b91d-093ee963d8cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.590302 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-jckvv"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.617503 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7f764a7a-b747-4b12-b91d-093ee963d8cb" (UID: "7f764a7a-b747-4b12-b91d-093ee963d8cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.626345 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d7849547c-j77g9"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.633364 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.633401 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.636654 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f764a7a-b747-4b12-b91d-093ee963d8cb" (UID: "7f764a7a-b747-4b12-b91d-093ee963d8cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.637099 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7f764a7a-b747-4b12-b91d-093ee963d8cb" (UID: "7f764a7a-b747-4b12-b91d-093ee963d8cb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.670180 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.727714 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.735248 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.735299 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f764a7a-b747-4b12-b91d-093ee963d8cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.757202 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lvcdn"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.779610 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-lvcdn"] Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.972807 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" event={"ID":"561700fb-a6d4-42e1-b339-f3dd062a504a","Type":"ContainerStarted","Data":"b1b9d9bc1ba77d572d43cc0b10870c0e405c47609e9db84f05cdc439f7fabd97"} Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.976108 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8","Type":"ContainerStarted","Data":"ae957ce0eb270cee8fe71608b36de9ac14a47258e80a4c59094aead06b818a6e"} Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.984552 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fnppb" event={"ID":"3cd84894-b4fa-481f-8856-678218f7bad7","Type":"ContainerStarted","Data":"5fe8c07896c62a9c99c258f7e2c8e1a57064244015771c1cac217a86a080bb78"} Dec 03 09:00:04 crc kubenswrapper[4573]: I1203 09:00:04.995915 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-srtdq" event={"ID":"8e8386e5-315f-4d6f-a560-e1a45777e339","Type":"ContainerStarted","Data":"5f52371526c50ea998b2a69364f5649fc0079b082b241c77c79fa578d7cffef1"} Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.003587 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c795868d9-6tqff" event={"ID":"5451c2e0-1168-4793-a1ba-c07db67f7b8a","Type":"ContainerStarted","Data":"bd1802e025c9269f68592366c9a601d66f03478681291035f68d9b4f2e06ed4c"} Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.029572 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" event={"ID":"1f1b04c5-a752-4557-a537-ad477d6d6280","Type":"ContainerStarted","Data":"a469fd84d82c8cd1438a1477c1d68221fb360047a50a0e0bb04e597e32b6a7c3"} Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.044339 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-kvbg6" event={"ID":"df6a51aa-75ab-4bdc-9e22-6ba714772ecf","Type":"ContainerStarted","Data":"55a8f18aea332a924ed619229167131088fb4f1445db489d59c16f8e04990e0f"} Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.064556 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-zwp5l" event={"ID":"7f764a7a-b747-4b12-b91d-093ee963d8cb","Type":"ContainerDied","Data":"74a15500ed350db7d03665de3338f8e3972b763c0d25fe46e6ca8a2b21adfb62"} Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.064662 4573 scope.go:117] "RemoveContainer" containerID="f89ebb241acb65b1941be33886abf87153532abd430f9e49e64e2b85561497e6" Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.064869 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-zwp5l" Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.081468 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d7849547c-j77g9" event={"ID":"e511768a-604b-441d-a363-e5c522b39647","Type":"ContainerStarted","Data":"76c3270f3928e48dd3b8ba0071033d0cc5c1c8aea6aa5cc4a5be62fe15ba44c1"} Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.196716 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-zwp5l"] Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.222047 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-zwp5l"] Dec 03 09:00:05 crc kubenswrapper[4573]: I1203 09:00:05.496589 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bdd77d6bf-c2wp4"] Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.083779 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06a5e660-c639-4897-8968-d7fabbba27f8" path="/var/lib/kubelet/pods/06a5e660-c639-4897-8968-d7fabbba27f8/volumes" Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.084690 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f764a7a-b747-4b12-b91d-093ee963d8cb" path="/var/lib/kubelet/pods/7f764a7a-b747-4b12-b91d-093ee963d8cb/volumes" Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.130867 4573 generic.go:334] "Generic (PLEG): container finished" podID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerID="47acf997deb639b2ea9e482a5ebdaa726ded6fcd513c9b477eb612e983e3de27" exitCode=0 Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.130944 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" event={"ID":"1f1b04c5-a752-4557-a537-ad477d6d6280","Type":"ContainerDied","Data":"47acf997deb639b2ea9e482a5ebdaa726ded6fcd513c9b477eb612e983e3de27"} Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.157030 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bdd77d6bf-c2wp4" event={"ID":"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea","Type":"ContainerStarted","Data":"73853110dff91a147cf6077bea4e52a64fb3cfcfdca339bfc098a6f6c08251a9"} Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.180206 4573 generic.go:334] "Generic (PLEG): container finished" podID="561700fb-a6d4-42e1-b339-f3dd062a504a" containerID="97bc403af90b64013e601cd9d66868509687cd01e94210a046eb56bdc4d9e1fa" exitCode=0 Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.180671 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" event={"ID":"561700fb-a6d4-42e1-b339-f3dd062a504a","Type":"ContainerDied","Data":"97bc403af90b64013e601cd9d66868509687cd01e94210a046eb56bdc4d9e1fa"} Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.240308 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-srtdq" event={"ID":"8e8386e5-315f-4d6f-a560-e1a45777e339","Type":"ContainerStarted","Data":"efd263f20f857341d85f4238f1a9435b948d198a4e7f1b02d40e17485eb33a8b"} Dec 03 09:00:06 crc kubenswrapper[4573]: I1203 09:00:06.296874 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-srtdq" podStartSLOduration=7.296834604 podStartE2EDuration="7.296834604s" podCreationTimestamp="2025-12-03 08:59:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:00:06.275771278 +0000 UTC m=+1326.844150537" watchObservedRunningTime="2025-12-03 09:00:06.296834604 +0000 UTC m=+1326.865213883" Dec 03 09:00:07 crc kubenswrapper[4573]: I1203 09:00:07.293726 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" event={"ID":"1f1b04c5-a752-4557-a537-ad477d6d6280","Type":"ContainerStarted","Data":"fbf6aa1c5cec0d6236795a987b1189668404b7ceae8aa0b0a0cfbdc073a0aaf7"} Dec 03 09:00:07 crc kubenswrapper[4573]: I1203 09:00:07.297716 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:07 crc kubenswrapper[4573]: I1203 09:00:07.421286 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" podStartSLOduration=8.421244157 podStartE2EDuration="8.421244157s" podCreationTimestamp="2025-12-03 08:59:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:00:07.414318177 +0000 UTC m=+1327.982697446" watchObservedRunningTime="2025-12-03 09:00:07.421244157 +0000 UTC m=+1327.989623416" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.286412 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.321062 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp8kw\" (UniqueName: \"kubernetes.io/projected/561700fb-a6d4-42e1-b339-f3dd062a504a-kube-api-access-fp8kw\") pod \"561700fb-a6d4-42e1-b339-f3dd062a504a\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.321352 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/561700fb-a6d4-42e1-b339-f3dd062a504a-secret-volume\") pod \"561700fb-a6d4-42e1-b339-f3dd062a504a\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.321414 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/561700fb-a6d4-42e1-b339-f3dd062a504a-config-volume\") pod \"561700fb-a6d4-42e1-b339-f3dd062a504a\" (UID: \"561700fb-a6d4-42e1-b339-f3dd062a504a\") " Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.322957 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/561700fb-a6d4-42e1-b339-f3dd062a504a-config-volume" (OuterVolumeSpecName: "config-volume") pod "561700fb-a6d4-42e1-b339-f3dd062a504a" (UID: "561700fb-a6d4-42e1-b339-f3dd062a504a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.349778 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561700fb-a6d4-42e1-b339-f3dd062a504a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "561700fb-a6d4-42e1-b339-f3dd062a504a" (UID: "561700fb-a6d4-42e1-b339-f3dd062a504a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.358997 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561700fb-a6d4-42e1-b339-f3dd062a504a-kube-api-access-fp8kw" (OuterVolumeSpecName: "kube-api-access-fp8kw") pod "561700fb-a6d4-42e1-b339-f3dd062a504a" (UID: "561700fb-a6d4-42e1-b339-f3dd062a504a"). InnerVolumeSpecName "kube-api-access-fp8kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.385228 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" event={"ID":"561700fb-a6d4-42e1-b339-f3dd062a504a","Type":"ContainerDied","Data":"b1b9d9bc1ba77d572d43cc0b10870c0e405c47609e9db84f05cdc439f7fabd97"} Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.385293 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1b9d9bc1ba77d572d43cc0b10870c0e405c47609e9db84f05cdc439f7fabd97" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.385833 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.435057 4573 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/561700fb-a6d4-42e1-b339-f3dd062a504a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.435135 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/561700fb-a6d4-42e1-b339-f3dd062a504a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:08 crc kubenswrapper[4573]: I1203 09:00:08.435150 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp8kw\" (UniqueName: \"kubernetes.io/projected/561700fb-a6d4-42e1-b339-f3dd062a504a-kube-api-access-fp8kw\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.522309 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d7849547c-j77g9"] Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.570631 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b8b5b6556-4zxxs"] Dec 03 09:00:10 crc kubenswrapper[4573]: E1203 09:00:10.571446 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f764a7a-b747-4b12-b91d-093ee963d8cb" containerName="init" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.572019 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f764a7a-b747-4b12-b91d-093ee963d8cb" containerName="init" Dec 03 09:00:10 crc kubenswrapper[4573]: E1203 09:00:10.572156 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561700fb-a6d4-42e1-b339-f3dd062a504a" containerName="collect-profiles" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.572235 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="561700fb-a6d4-42e1-b339-f3dd062a504a" containerName="collect-profiles" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.572491 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="561700fb-a6d4-42e1-b339-f3dd062a504a" containerName="collect-profiles" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.572569 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f764a7a-b747-4b12-b91d-093ee963d8cb" containerName="init" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.581296 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.594580 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.602461 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b8b5b6556-4zxxs"] Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.689024 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bdd77d6bf-c2wp4"] Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.712728 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-tls-certs\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.712821 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-secret-key\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.712897 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-577vk\" (UniqueName: \"kubernetes.io/projected/8c69ceb8-3d0e-4920-be28-36799e264f96-kube-api-access-577vk\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.712943 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-config-data\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.713033 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-combined-ca-bundle\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.713092 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c69ceb8-3d0e-4920-be28-36799e264f96-logs\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.713344 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-scripts\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.730871 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66f99c579b-gh7sh"] Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.735367 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.790432 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66f99c579b-gh7sh"] Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.815035 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-combined-ca-bundle\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.815118 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c69ceb8-3d0e-4920-be28-36799e264f96-logs\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.815199 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-scripts\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.815239 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-tls-certs\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.815269 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-secret-key\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.815314 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-577vk\" (UniqueName: \"kubernetes.io/projected/8c69ceb8-3d0e-4920-be28-36799e264f96-kube-api-access-577vk\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.815347 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-config-data\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.816411 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-scripts\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.816804 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c69ceb8-3d0e-4920-be28-36799e264f96-logs\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.817117 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-config-data\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.827904 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-combined-ca-bundle\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.835442 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-tls-certs\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.837179 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-577vk\" (UniqueName: \"kubernetes.io/projected/8c69ceb8-3d0e-4920-be28-36799e264f96-kube-api-access-577vk\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.850187 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-secret-key\") pod \"horizon-7b8b5b6556-4zxxs\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.918136 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.918216 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-combined-ca-bundle\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.918426 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-horizon-secret-key\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.918464 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b78e2453-adc3-4139-97c5-9f8a977df1ee-logs\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.918494 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b78e2453-adc3-4139-97c5-9f8a977df1ee-scripts\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.918517 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b78e2453-adc3-4139-97c5-9f8a977df1ee-config-data\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.918560 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-horizon-tls-certs\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:10 crc kubenswrapper[4573]: I1203 09:00:10.918596 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pndh7\" (UniqueName: \"kubernetes.io/projected/b78e2453-adc3-4139-97c5-9f8a977df1ee-kube-api-access-pndh7\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.014123 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.020501 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-combined-ca-bundle\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.020677 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-horizon-secret-key\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.020711 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b78e2453-adc3-4139-97c5-9f8a977df1ee-logs\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.020796 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b78e2453-adc3-4139-97c5-9f8a977df1ee-scripts\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.020838 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b78e2453-adc3-4139-97c5-9f8a977df1ee-config-data\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.020879 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-horizon-tls-certs\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.020930 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pndh7\" (UniqueName: \"kubernetes.io/projected/b78e2453-adc3-4139-97c5-9f8a977df1ee-kube-api-access-pndh7\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.022209 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b78e2453-adc3-4139-97c5-9f8a977df1ee-logs\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.022373 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b78e2453-adc3-4139-97c5-9f8a977df1ee-scripts\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.023599 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b78e2453-adc3-4139-97c5-9f8a977df1ee-config-data\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.046956 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-horizon-secret-key\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.047096 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-horizon-tls-certs\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.047585 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78e2453-adc3-4139-97c5-9f8a977df1ee-combined-ca-bundle\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.069186 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pndh7\" (UniqueName: \"kubernetes.io/projected/b78e2453-adc3-4139-97c5-9f8a977df1ee-kube-api-access-pndh7\") pod \"horizon-66f99c579b-gh7sh\" (UID: \"b78e2453-adc3-4139-97c5-9f8a977df1ee\") " pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.070469 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.131671 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-cl5cv"] Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.132057 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" podUID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" containerName="dnsmasq-dns" containerID="cri-o://6ef8e5934c94595f86df204e3ac8183e89565374ccead7c035e8695fe6a1b12c" gracePeriod=10 Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.613872 4573 generic.go:334] "Generic (PLEG): container finished" podID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" containerID="6ef8e5934c94595f86df204e3ac8183e89565374ccead7c035e8695fe6a1b12c" exitCode=0 Dec 03 09:00:11 crc kubenswrapper[4573]: I1203 09:00:11.619163 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" event={"ID":"d3e18b14-1f6a-4cb6-af58-398d0e393ffe","Type":"ContainerDied","Data":"6ef8e5934c94595f86df204e3ac8183e89565374ccead7c035e8695fe6a1b12c"} Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.460794 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66f99c579b-gh7sh"] Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.671475 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" event={"ID":"d3e18b14-1f6a-4cb6-af58-398d0e393ffe","Type":"ContainerDied","Data":"104b116b43f367c73c2c1ec234bbcf52f46ca85bdda9360dab38b8e34d889688"} Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.672187 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="104b116b43f367c73c2c1ec234bbcf52f46ca85bdda9360dab38b8e34d889688" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.671539 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.680134 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66f99c579b-gh7sh" event={"ID":"b78e2453-adc3-4139-97c5-9f8a977df1ee","Type":"ContainerStarted","Data":"3e2ea1a082078a8aa04cb4a95b5d9a16bdcd42de18a8ac2ae7b84e793b846321"} Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.696205 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b8b5b6556-4zxxs"] Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.747609 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-config\") pod \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.747740 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-dns-svc\") pod \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.747858 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-sb\") pod \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.747936 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltxmg\" (UniqueName: \"kubernetes.io/projected/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-kube-api-access-ltxmg\") pod \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.747970 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-nb\") pod \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\" (UID: \"d3e18b14-1f6a-4cb6-af58-398d0e393ffe\") " Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.779357 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-kube-api-access-ltxmg" (OuterVolumeSpecName: "kube-api-access-ltxmg") pod "d3e18b14-1f6a-4cb6-af58-398d0e393ffe" (UID: "d3e18b14-1f6a-4cb6-af58-398d0e393ffe"). InnerVolumeSpecName "kube-api-access-ltxmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.815075 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3e18b14-1f6a-4cb6-af58-398d0e393ffe" (UID: "d3e18b14-1f6a-4cb6-af58-398d0e393ffe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.862604 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltxmg\" (UniqueName: \"kubernetes.io/projected/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-kube-api-access-ltxmg\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.862643 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.895290 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d3e18b14-1f6a-4cb6-af58-398d0e393ffe" (UID: "d3e18b14-1f6a-4cb6-af58-398d0e393ffe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.934312 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-config" (OuterVolumeSpecName: "config") pod "d3e18b14-1f6a-4cb6-af58-398d0e393ffe" (UID: "d3e18b14-1f6a-4cb6-af58-398d0e393ffe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.964643 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.972195 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:12 crc kubenswrapper[4573]: I1203 09:00:12.981349 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3e18b14-1f6a-4cb6-af58-398d0e393ffe" (UID: "d3e18b14-1f6a-4cb6-af58-398d0e393ffe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:13 crc kubenswrapper[4573]: I1203 09:00:13.074807 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3e18b14-1f6a-4cb6-af58-398d0e393ffe-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:13 crc kubenswrapper[4573]: I1203 09:00:13.716917 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerStarted","Data":"4dc9c0026737b9e06e10ab08990fb2f6fac86fe8064e66cd1d53ce7d88f5fd13"} Dec 03 09:00:13 crc kubenswrapper[4573]: I1203 09:00:13.724646 4573 generic.go:334] "Generic (PLEG): container finished" podID="a051e848-8cfe-4ea3-bed8-721665a5a37b" containerID="c5991229e5c00013cea09dc1f612403384de8e841bcf82e821749bd6642ea1a4" exitCode=0 Dec 03 09:00:13 crc kubenswrapper[4573]: I1203 09:00:13.724844 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-cl5cv" Dec 03 09:00:13 crc kubenswrapper[4573]: I1203 09:00:13.728824 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v556f" event={"ID":"a051e848-8cfe-4ea3-bed8-721665a5a37b","Type":"ContainerDied","Data":"c5991229e5c00013cea09dc1f612403384de8e841bcf82e821749bd6642ea1a4"} Dec 03 09:00:13 crc kubenswrapper[4573]: I1203 09:00:13.792860 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-cl5cv"] Dec 03 09:00:13 crc kubenswrapper[4573]: I1203 09:00:13.803411 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-cl5cv"] Dec 03 09:00:14 crc kubenswrapper[4573]: I1203 09:00:14.056378 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" path="/var/lib/kubelet/pods/d3e18b14-1f6a-4cb6-af58-398d0e393ffe/volumes" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.485375 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v556f" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.536271 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-config-data\") pod \"a051e848-8cfe-4ea3-bed8-721665a5a37b\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.536469 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-combined-ca-bundle\") pod \"a051e848-8cfe-4ea3-bed8-721665a5a37b\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.536542 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-credential-keys\") pod \"a051e848-8cfe-4ea3-bed8-721665a5a37b\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.536562 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-fernet-keys\") pod \"a051e848-8cfe-4ea3-bed8-721665a5a37b\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.536611 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-scripts\") pod \"a051e848-8cfe-4ea3-bed8-721665a5a37b\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.536668 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56tsq\" (UniqueName: \"kubernetes.io/projected/a051e848-8cfe-4ea3-bed8-721665a5a37b-kube-api-access-56tsq\") pod \"a051e848-8cfe-4ea3-bed8-721665a5a37b\" (UID: \"a051e848-8cfe-4ea3-bed8-721665a5a37b\") " Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.545792 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a051e848-8cfe-4ea3-bed8-721665a5a37b" (UID: "a051e848-8cfe-4ea3-bed8-721665a5a37b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.560109 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-scripts" (OuterVolumeSpecName: "scripts") pod "a051e848-8cfe-4ea3-bed8-721665a5a37b" (UID: "a051e848-8cfe-4ea3-bed8-721665a5a37b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.560314 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a051e848-8cfe-4ea3-bed8-721665a5a37b-kube-api-access-56tsq" (OuterVolumeSpecName: "kube-api-access-56tsq") pod "a051e848-8cfe-4ea3-bed8-721665a5a37b" (UID: "a051e848-8cfe-4ea3-bed8-721665a5a37b"). InnerVolumeSpecName "kube-api-access-56tsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.578962 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a051e848-8cfe-4ea3-bed8-721665a5a37b" (UID: "a051e848-8cfe-4ea3-bed8-721665a5a37b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.579893 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-config-data" (OuterVolumeSpecName: "config-data") pod "a051e848-8cfe-4ea3-bed8-721665a5a37b" (UID: "a051e848-8cfe-4ea3-bed8-721665a5a37b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.594899 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a051e848-8cfe-4ea3-bed8-721665a5a37b" (UID: "a051e848-8cfe-4ea3-bed8-721665a5a37b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.639601 4573 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.639647 4573 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.639659 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.639671 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56tsq\" (UniqueName: \"kubernetes.io/projected/a051e848-8cfe-4ea3-bed8-721665a5a37b-kube-api-access-56tsq\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.639710 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.639719 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a051e848-8cfe-4ea3-bed8-721665a5a37b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.813177 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v556f" event={"ID":"a051e848-8cfe-4ea3-bed8-721665a5a37b","Type":"ContainerDied","Data":"ddb53d54270fd9eb922defe011498666c77a7d1dbaba4ce470434f5f44bd27f7"} Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.813267 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddb53d54270fd9eb922defe011498666c77a7d1dbaba4ce470434f5f44bd27f7" Dec 03 09:00:18 crc kubenswrapper[4573]: I1203 09:00:18.813306 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v556f" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.614726 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v556f"] Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.624576 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v556f"] Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.707244 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-m4vz4"] Dec 03 09:00:19 crc kubenswrapper[4573]: E1203 09:00:19.709033 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" containerName="init" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.709250 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" containerName="init" Dec 03 09:00:19 crc kubenswrapper[4573]: E1203 09:00:19.709331 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a051e848-8cfe-4ea3-bed8-721665a5a37b" containerName="keystone-bootstrap" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.709399 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="a051e848-8cfe-4ea3-bed8-721665a5a37b" containerName="keystone-bootstrap" Dec 03 09:00:19 crc kubenswrapper[4573]: E1203 09:00:19.709459 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" containerName="dnsmasq-dns" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.709513 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" containerName="dnsmasq-dns" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.709809 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="a051e848-8cfe-4ea3-bed8-721665a5a37b" containerName="keystone-bootstrap" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.711628 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e18b14-1f6a-4cb6-af58-398d0e393ffe" containerName="dnsmasq-dns" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.712923 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.720272 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5wmq4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.720655 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.720873 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.721100 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.728690 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m4vz4"] Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.729621 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.794012 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpnxr\" (UniqueName: \"kubernetes.io/projected/e28fb330-9e1a-4480-9a2d-b059d8e1576d-kube-api-access-xpnxr\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.794666 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-scripts\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.794722 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-combined-ca-bundle\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.795023 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-credential-keys\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.795124 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-config-data\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.795193 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-fernet-keys\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.900521 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-scripts\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.900594 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-combined-ca-bundle\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.900667 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-credential-keys\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.900688 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-config-data\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.900711 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-fernet-keys\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.900759 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpnxr\" (UniqueName: \"kubernetes.io/projected/e28fb330-9e1a-4480-9a2d-b059d8e1576d-kube-api-access-xpnxr\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.907158 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-config-data\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.908768 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-fernet-keys\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.911719 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-credential-keys\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.912759 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-combined-ca-bundle\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.912899 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-scripts\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:19 crc kubenswrapper[4573]: I1203 09:00:19.930029 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpnxr\" (UniqueName: \"kubernetes.io/projected/e28fb330-9e1a-4480-9a2d-b059d8e1576d-kube-api-access-xpnxr\") pod \"keystone-bootstrap-m4vz4\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:20 crc kubenswrapper[4573]: I1203 09:00:20.049303 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:00:20 crc kubenswrapper[4573]: I1203 09:00:20.052941 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a051e848-8cfe-4ea3-bed8-721665a5a37b" path="/var/lib/kubelet/pods/a051e848-8cfe-4ea3-bed8-721665a5a37b/volumes" Dec 03 09:00:25 crc kubenswrapper[4573]: I1203 09:00:25.922839 4573 generic.go:334] "Generic (PLEG): container finished" podID="de65a9c9-bef6-42e3-b256-f94d4af37c96" containerID="84dcfceb254e96342219e95412e54d29973b82e2655f344fb62ea90137bce299" exitCode=0 Dec 03 09:00:25 crc kubenswrapper[4573]: I1203 09:00:25.923143 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8t69f" event={"ID":"de65a9c9-bef6-42e3-b256-f94d4af37c96","Type":"ContainerDied","Data":"84dcfceb254e96342219e95412e54d29973b82e2655f344fb62ea90137bce299"} Dec 03 09:00:26 crc kubenswrapper[4573]: E1203 09:00:26.688679 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 09:00:26 crc kubenswrapper[4573]: E1203 09:00:26.689429 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n584h8dh97h564h699hd8h68bh87h57ch595h5dbh6fh5c4h5f4h5fdh544h67fh57bh674h7h5d8h5b8h5f6h65chbfh5f5h675hfch655hf9h585h595q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7kkcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6c795868d9-6tqff_openstack(5451c2e0-1168-4793-a1ba-c07db67f7b8a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:26 crc kubenswrapper[4573]: E1203 09:00:26.706276 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6c795868d9-6tqff" podUID="5451c2e0-1168-4793-a1ba-c07db67f7b8a" Dec 03 09:00:30 crc kubenswrapper[4573]: E1203 09:00:30.818483 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 09:00:30 crc kubenswrapper[4573]: E1203 09:00:30.819518 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n675h547hc9h586hbfhfbhb6h594h68fh544h77h56bh66h98h5cch58ch675h645h5c4hbdh5b4h578h584h5bch666h66fh648h568h7ch84hcch68q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rb5sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:30 crc kubenswrapper[4573]: E1203 09:00:30.836398 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 09:00:30 crc kubenswrapper[4573]: E1203 09:00:30.836670 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n58bh699h64fh5hbch5f4h645hf4h5fch5b4h5f7h8dhffh687h57h69h7dh5fbh9fh5bch5dh5d7h5b7h556h9bh8bh77hcfh87h59bh556h546q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c7f6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5bdd77d6bf-c2wp4_openstack(b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:30 crc kubenswrapper[4573]: E1203 09:00:30.843469 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5bdd77d6bf-c2wp4" podUID="b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea" Dec 03 09:00:30 crc kubenswrapper[4573]: E1203 09:00:30.852573 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 09:00:30 crc kubenswrapper[4573]: E1203 09:00:30.852900 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8h55fh588hb5h657hd5h677h645h7h558h694h555h99h9h55bhddh7fh688h64h66ch58bh67h5bch688h5cdh564h565h5f5h5dbh5bbh644h5dbq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pndh7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-66f99c579b-gh7sh_openstack(b78e2453-adc3-4139-97c5-9f8a977df1ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:30 crc kubenswrapper[4573]: E1203 09:00:30.855008 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" Dec 03 09:00:30 crc kubenswrapper[4573]: I1203 09:00:30.936804 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8t69f" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.000289 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8t69f" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.000377 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8t69f" event={"ID":"de65a9c9-bef6-42e3-b256-f94d4af37c96","Type":"ContainerDied","Data":"07b8fd8767b51755d0dfcfb57e94ba1e1a90c6c683f8b62ebb151ec9206f3d61"} Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.000460 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07b8fd8767b51755d0dfcfb57e94ba1e1a90c6c683f8b62ebb151ec9206f3d61" Dec 03 09:00:31 crc kubenswrapper[4573]: E1203 09:00:31.010314 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.090389 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-config-data\") pod \"de65a9c9-bef6-42e3-b256-f94d4af37c96\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.090468 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c45k\" (UniqueName: \"kubernetes.io/projected/de65a9c9-bef6-42e3-b256-f94d4af37c96-kube-api-access-9c45k\") pod \"de65a9c9-bef6-42e3-b256-f94d4af37c96\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.090796 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-combined-ca-bundle\") pod \"de65a9c9-bef6-42e3-b256-f94d4af37c96\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.090825 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-db-sync-config-data\") pod \"de65a9c9-bef6-42e3-b256-f94d4af37c96\" (UID: \"de65a9c9-bef6-42e3-b256-f94d4af37c96\") " Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.099478 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de65a9c9-bef6-42e3-b256-f94d4af37c96-kube-api-access-9c45k" (OuterVolumeSpecName: "kube-api-access-9c45k") pod "de65a9c9-bef6-42e3-b256-f94d4af37c96" (UID: "de65a9c9-bef6-42e3-b256-f94d4af37c96"). InnerVolumeSpecName "kube-api-access-9c45k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.103848 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "de65a9c9-bef6-42e3-b256-f94d4af37c96" (UID: "de65a9c9-bef6-42e3-b256-f94d4af37c96"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.137982 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de65a9c9-bef6-42e3-b256-f94d4af37c96" (UID: "de65a9c9-bef6-42e3-b256-f94d4af37c96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.189615 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-config-data" (OuterVolumeSpecName: "config-data") pod "de65a9c9-bef6-42e3-b256-f94d4af37c96" (UID: "de65a9c9-bef6-42e3-b256-f94d4af37c96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.199818 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.199863 4573 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.199876 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de65a9c9-bef6-42e3-b256-f94d4af37c96-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:31 crc kubenswrapper[4573]: I1203 09:00:31.199885 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c45k\" (UniqueName: \"kubernetes.io/projected/de65a9c9-bef6-42e3-b256-f94d4af37c96-kube-api-access-9c45k\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.625200 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-fd87c"] Dec 03 09:00:32 crc kubenswrapper[4573]: E1203 09:00:32.628057 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de65a9c9-bef6-42e3-b256-f94d4af37c96" containerName="glance-db-sync" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.628082 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="de65a9c9-bef6-42e3-b256-f94d4af37c96" containerName="glance-db-sync" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.628334 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="de65a9c9-bef6-42e3-b256-f94d4af37c96" containerName="glance-db-sync" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.632543 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.664602 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-config\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.664666 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.664687 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.664791 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.664817 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.664838 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njkwv\" (UniqueName: \"kubernetes.io/projected/032f4b3b-1037-4237-ba70-77d3f9753d9e-kube-api-access-njkwv\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.678464 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-fd87c"] Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.771563 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-config\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.771638 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.771659 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.771770 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.771801 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.771823 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njkwv\" (UniqueName: \"kubernetes.io/projected/032f4b3b-1037-4237-ba70-77d3f9753d9e-kube-api-access-njkwv\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.772709 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-config\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.773610 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.773621 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.775572 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.775987 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:32 crc kubenswrapper[4573]: I1203 09:00:32.819846 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njkwv\" (UniqueName: \"kubernetes.io/projected/032f4b3b-1037-4237-ba70-77d3f9753d9e-kube-api-access-njkwv\") pod \"dnsmasq-dns-56df8fb6b7-fd87c\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.037701 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.349747 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.352592 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.356281 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.358739 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.360808 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gx6qj" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.385344 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.398969 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.399068 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.399122 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.399152 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.399245 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gf6m\" (UniqueName: \"kubernetes.io/projected/8016ac24-adaf-481c-8ad2-6016bc588fa9-kube-api-access-5gf6m\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.399284 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.399304 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-logs\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.500453 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gf6m\" (UniqueName: \"kubernetes.io/projected/8016ac24-adaf-481c-8ad2-6016bc588fa9-kube-api-access-5gf6m\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.500761 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.500906 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-logs\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.501032 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.501197 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.501334 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.501650 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.502257 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.501979 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-logs\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.501813 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.512564 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.514428 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-scripts\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.520462 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-config-data\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.539946 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gf6m\" (UniqueName: \"kubernetes.io/projected/8016ac24-adaf-481c-8ad2-6016bc588fa9-kube-api-access-5gf6m\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.623489 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.683828 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.980502 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.982804 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:00:33 crc kubenswrapper[4573]: I1203 09:00:33.987357 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.021054 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.129550 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.129661 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-logs\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.129807 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.129896 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.129926 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch7ws\" (UniqueName: \"kubernetes.io/projected/53de8a70-c359-4b86-a783-cf1815b73d68-kube-api-access-ch7ws\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.129986 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.130550 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.232316 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.232422 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.233708 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-logs\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.233760 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.233793 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.233813 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch7ws\" (UniqueName: \"kubernetes.io/projected/53de8a70-c359-4b86-a783-cf1815b73d68-kube-api-access-ch7ws\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.233849 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.234337 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.234627 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-logs\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.237519 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.262324 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.274128 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.278165 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch7ws\" (UniqueName: \"kubernetes.io/projected/53de8a70-c359-4b86-a783-cf1815b73d68-kube-api-access-ch7ws\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.287812 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.368571 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:00:34 crc kubenswrapper[4573]: I1203 09:00:34.669310 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:00:35 crc kubenswrapper[4573]: E1203 09:00:35.664511 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 09:00:35 crc kubenswrapper[4573]: E1203 09:00:35.665035 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5f9h5b6h584h679h565h68fhb6h7fh64bh58ch545h675h675h5c6h64bhd6h58ch87h548h674h55dh65chbh699hffh7ch5h4h55bh64fh5cdh546q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-577vk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7b8b5b6556-4zxxs_openstack(8c69ceb8-3d0e-4920-be28-36799e264f96): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:35 crc kubenswrapper[4573]: E1203 09:00:35.675364 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" Dec 03 09:00:36 crc kubenswrapper[4573]: E1203 09:00:36.073158 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" Dec 03 09:00:36 crc kubenswrapper[4573]: I1203 09:00:36.496242 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:00:36 crc kubenswrapper[4573]: I1203 09:00:36.613314 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.328245 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c795868d9-6tqff" Dec 03 09:00:48 crc kubenswrapper[4573]: E1203 09:00:48.350545 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 09:00:48 crc kubenswrapper[4573]: E1203 09:00:48.350824 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n67ch8ch544h5dch56dhf9h56bh597hcdh68ch56bh598h566h646h5dch569hd4h5d9h7h7h586h84h667h58ch8h5ch68h4h657h64bh5fh677q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-84ddg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7d7849547c-j77g9_openstack(e511768a-604b-441d-a363-e5c522b39647): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:48 crc kubenswrapper[4573]: E1203 09:00:48.364024 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d7849547c-j77g9" podUID="e511768a-604b-441d-a363-e5c522b39647" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.449001 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5451c2e0-1168-4793-a1ba-c07db67f7b8a-horizon-secret-key\") pod \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.449502 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-scripts\") pod \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.449576 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-config-data\") pod \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.449616 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kkcl\" (UniqueName: \"kubernetes.io/projected/5451c2e0-1168-4793-a1ba-c07db67f7b8a-kube-api-access-7kkcl\") pod \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.449802 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5451c2e0-1168-4793-a1ba-c07db67f7b8a-logs\") pod \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\" (UID: \"5451c2e0-1168-4793-a1ba-c07db67f7b8a\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.451336 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-scripts" (OuterVolumeSpecName: "scripts") pod "5451c2e0-1168-4793-a1ba-c07db67f7b8a" (UID: "5451c2e0-1168-4793-a1ba-c07db67f7b8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.451618 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5451c2e0-1168-4793-a1ba-c07db67f7b8a-logs" (OuterVolumeSpecName: "logs") pod "5451c2e0-1168-4793-a1ba-c07db67f7b8a" (UID: "5451c2e0-1168-4793-a1ba-c07db67f7b8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.451726 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-config-data" (OuterVolumeSpecName: "config-data") pod "5451c2e0-1168-4793-a1ba-c07db67f7b8a" (UID: "5451c2e0-1168-4793-a1ba-c07db67f7b8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.463157 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5451c2e0-1168-4793-a1ba-c07db67f7b8a-kube-api-access-7kkcl" (OuterVolumeSpecName: "kube-api-access-7kkcl") pod "5451c2e0-1168-4793-a1ba-c07db67f7b8a" (UID: "5451c2e0-1168-4793-a1ba-c07db67f7b8a"). InnerVolumeSpecName "kube-api-access-7kkcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.467951 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5451c2e0-1168-4793-a1ba-c07db67f7b8a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "5451c2e0-1168-4793-a1ba-c07db67f7b8a" (UID: "5451c2e0-1168-4793-a1ba-c07db67f7b8a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.512353 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.551155 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-config-data\") pod \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.551237 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-scripts\") pod \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.551379 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-horizon-secret-key\") pod \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.552158 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-config-data" (OuterVolumeSpecName: "config-data") pod "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea" (UID: "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.552286 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-logs\") pod \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.552491 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7f6n\" (UniqueName: \"kubernetes.io/projected/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-kube-api-access-c7f6n\") pod \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\" (UID: \"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea\") " Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.552264 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-scripts" (OuterVolumeSpecName: "scripts") pod "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea" (UID: "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.552507 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-logs" (OuterVolumeSpecName: "logs") pod "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea" (UID: "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.553015 4573 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5451c2e0-1168-4793-a1ba-c07db67f7b8a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.553065 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.553084 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5451c2e0-1168-4793-a1ba-c07db67f7b8a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.553096 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kkcl\" (UniqueName: \"kubernetes.io/projected/5451c2e0-1168-4793-a1ba-c07db67f7b8a-kube-api-access-7kkcl\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.553106 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.553115 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5451c2e0-1168-4793-a1ba-c07db67f7b8a-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.553125 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.553134 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.554923 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea" (UID: "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.557448 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-kube-api-access-c7f6n" (OuterVolumeSpecName: "kube-api-access-c7f6n") pod "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea" (UID: "b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea"). InnerVolumeSpecName "kube-api-access-c7f6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.655093 4573 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:48 crc kubenswrapper[4573]: I1203 09:00:48.655154 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7f6n\" (UniqueName: \"kubernetes.io/projected/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea-kube-api-access-c7f6n\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:49 crc kubenswrapper[4573]: E1203 09:00:49.142971 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 03 09:00:49 crc kubenswrapper[4573]: E1203 09:00:49.143202 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xxl4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-fnppb_openstack(3cd84894-b4fa-481f-8856-678218f7bad7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:49 crc kubenswrapper[4573]: E1203 09:00:49.144608 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-fnppb" podUID="3cd84894-b4fa-481f-8856-678218f7bad7" Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.228997 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bdd77d6bf-c2wp4" event={"ID":"b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea","Type":"ContainerDied","Data":"73853110dff91a147cf6077bea4e52a64fb3cfcfdca339bfc098a6f6c08251a9"} Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.229137 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bdd77d6bf-c2wp4" Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.235442 4573 generic.go:334] "Generic (PLEG): container finished" podID="8e8386e5-315f-4d6f-a560-e1a45777e339" containerID="efd263f20f857341d85f4238f1a9435b948d198a4e7f1b02d40e17485eb33a8b" exitCode=0 Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.235498 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-srtdq" event={"ID":"8e8386e5-315f-4d6f-a560-e1a45777e339","Type":"ContainerDied","Data":"efd263f20f857341d85f4238f1a9435b948d198a4e7f1b02d40e17485eb33a8b"} Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.244493 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c795868d9-6tqff" event={"ID":"5451c2e0-1168-4793-a1ba-c07db67f7b8a","Type":"ContainerDied","Data":"bd1802e025c9269f68592366c9a601d66f03478681291035f68d9b4f2e06ed4c"} Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.244541 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c795868d9-6tqff" Dec 03 09:00:49 crc kubenswrapper[4573]: E1203 09:00:49.248924 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-fnppb" podUID="3cd84894-b4fa-481f-8856-678218f7bad7" Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.454313 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c795868d9-6tqff"] Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.468315 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6c795868d9-6tqff"] Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.489167 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bdd77d6bf-c2wp4"] Dec 03 09:00:49 crc kubenswrapper[4573]: I1203 09:00:49.501064 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bdd77d6bf-c2wp4"] Dec 03 09:00:50 crc kubenswrapper[4573]: I1203 09:00:50.057335 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5451c2e0-1168-4793-a1ba-c07db67f7b8a" path="/var/lib/kubelet/pods/5451c2e0-1168-4793-a1ba-c07db67f7b8a/volumes" Dec 03 09:00:50 crc kubenswrapper[4573]: I1203 09:00:50.060283 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea" path="/var/lib/kubelet/pods/b1cdffa7-e68f-4a7f-8eb2-b26ece42ddea/volumes" Dec 03 09:00:51 crc kubenswrapper[4573]: E1203 09:00:51.476104 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 09:00:51 crc kubenswrapper[4573]: E1203 09:00:51.476907 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vwfgd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-chs2b_openstack(5c73103b-e4fd-4a54-ad2c-021db8042b13): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:51 crc kubenswrapper[4573]: E1203 09:00:51.478114 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-chs2b" podUID="5c73103b-e4fd-4a54-ad2c-021db8042b13" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.043948 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m4vz4"] Dec 03 09:00:52 crc kubenswrapper[4573]: E1203 09:00:52.048182 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified" Dec 03 09:00:52 crc kubenswrapper[4573]: E1203 09:00:52.048439 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-notification-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n675h547hc9h586hbfhfbhb6h594h68fh544h77h56bh66h98h5cch58ch675h645h5c4hbdh5b4h578h584h5bch666h66fh648h568h7ch84hcch68q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-notification-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rb5sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/notificationhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.225220 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.274824 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84ddg\" (UniqueName: \"kubernetes.io/projected/e511768a-604b-441d-a363-e5c522b39647-kube-api-access-84ddg\") pod \"e511768a-604b-441d-a363-e5c522b39647\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.275018 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-scripts\") pod \"e511768a-604b-441d-a363-e5c522b39647\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.275117 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e511768a-604b-441d-a363-e5c522b39647-logs\") pod \"e511768a-604b-441d-a363-e5c522b39647\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.275271 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-config-data\") pod \"e511768a-604b-441d-a363-e5c522b39647\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.275304 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e511768a-604b-441d-a363-e5c522b39647-horizon-secret-key\") pod \"e511768a-604b-441d-a363-e5c522b39647\" (UID: \"e511768a-604b-441d-a363-e5c522b39647\") " Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.275693 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-scripts" (OuterVolumeSpecName: "scripts") pod "e511768a-604b-441d-a363-e5c522b39647" (UID: "e511768a-604b-441d-a363-e5c522b39647"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.276091 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e511768a-604b-441d-a363-e5c522b39647-logs" (OuterVolumeSpecName: "logs") pod "e511768a-604b-441d-a363-e5c522b39647" (UID: "e511768a-604b-441d-a363-e5c522b39647"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.276552 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-config-data" (OuterVolumeSpecName: "config-data") pod "e511768a-604b-441d-a363-e5c522b39647" (UID: "e511768a-604b-441d-a363-e5c522b39647"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.284731 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e511768a-604b-441d-a363-e5c522b39647-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e511768a-604b-441d-a363-e5c522b39647" (UID: "e511768a-604b-441d-a363-e5c522b39647"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.284778 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e511768a-604b-441d-a363-e5c522b39647-kube-api-access-84ddg" (OuterVolumeSpecName: "kube-api-access-84ddg") pod "e511768a-604b-441d-a363-e5c522b39647" (UID: "e511768a-604b-441d-a363-e5c522b39647"). InnerVolumeSpecName "kube-api-access-84ddg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.286997 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-srtdq" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.377859 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-combined-ca-bundle\") pod \"8e8386e5-315f-4d6f-a560-e1a45777e339\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.378090 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nxhw\" (UniqueName: \"kubernetes.io/projected/8e8386e5-315f-4d6f-a560-e1a45777e339-kube-api-access-5nxhw\") pod \"8e8386e5-315f-4d6f-a560-e1a45777e339\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.378140 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-config\") pod \"8e8386e5-315f-4d6f-a560-e1a45777e339\" (UID: \"8e8386e5-315f-4d6f-a560-e1a45777e339\") " Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.378916 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84ddg\" (UniqueName: \"kubernetes.io/projected/e511768a-604b-441d-a363-e5c522b39647-kube-api-access-84ddg\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.378941 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.378954 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e511768a-604b-441d-a363-e5c522b39647-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.378965 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e511768a-604b-441d-a363-e5c522b39647-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.378975 4573 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e511768a-604b-441d-a363-e5c522b39647-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.382771 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m4vz4" event={"ID":"e28fb330-9e1a-4480-9a2d-b059d8e1576d","Type":"ContainerStarted","Data":"7f58726d132b1790abcb1296dac03d02436caae1e6e9b42653d6e1c3ea7740b7"} Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.390337 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e8386e5-315f-4d6f-a560-e1a45777e339-kube-api-access-5nxhw" (OuterVolumeSpecName: "kube-api-access-5nxhw") pod "8e8386e5-315f-4d6f-a560-e1a45777e339" (UID: "8e8386e5-315f-4d6f-a560-e1a45777e339"). InnerVolumeSpecName "kube-api-access-5nxhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.426414 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-srtdq" event={"ID":"8e8386e5-315f-4d6f-a560-e1a45777e339","Type":"ContainerDied","Data":"5f52371526c50ea998b2a69364f5649fc0079b082b241c77c79fa578d7cffef1"} Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.426475 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f52371526c50ea998b2a69364f5649fc0079b082b241c77c79fa578d7cffef1" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.426551 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-srtdq" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.431397 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e8386e5-315f-4d6f-a560-e1a45777e339" (UID: "8e8386e5-315f-4d6f-a560-e1a45777e339"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.435456 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d7849547c-j77g9" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.436152 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d7849547c-j77g9" event={"ID":"e511768a-604b-441d-a363-e5c522b39647","Type":"ContainerDied","Data":"76c3270f3928e48dd3b8ba0071033d0cc5c1c8aea6aa5cc4a5be62fe15ba44c1"} Dec 03 09:00:52 crc kubenswrapper[4573]: E1203 09:00:52.437101 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-chs2b" podUID="5c73103b-e4fd-4a54-ad2c-021db8042b13" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.522602 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.523223 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nxhw\" (UniqueName: \"kubernetes.io/projected/8e8386e5-315f-4d6f-a560-e1a45777e339-kube-api-access-5nxhw\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.665641 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d7849547c-j77g9"] Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.689121 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d7849547c-j77g9"] Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.753017 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-config" (OuterVolumeSpecName: "config") pod "8e8386e5-315f-4d6f-a560-e1a45777e339" (UID: "8e8386e5-315f-4d6f-a560-e1a45777e339"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.754839 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-fd87c"] Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.756751 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e8386e5-315f-4d6f-a560-e1a45777e339-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:52 crc kubenswrapper[4573]: I1203 09:00:52.945898 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:00:52 crc kubenswrapper[4573]: W1203 09:00:52.948679 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53de8a70_c359_4b86_a783_cf1815b73d68.slice/crio-c4c68bcf7251acc1034639d03065ae89fd34ee08fa1c5a732eac7d35b50cd68f WatchSource:0}: Error finding container c4c68bcf7251acc1034639d03065ae89fd34ee08fa1c5a732eac7d35b50cd68f: Status 404 returned error can't find the container with id c4c68bcf7251acc1034639d03065ae89fd34ee08fa1c5a732eac7d35b50cd68f Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.143083 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:00:53 crc kubenswrapper[4573]: W1203 09:00:53.161222 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8016ac24_adaf_481c_8ad2_6016bc588fa9.slice/crio-3948f69bf2c48d85498b2f26e26361492896446a09bffa30a4abdfe278bbdf8c WatchSource:0}: Error finding container 3948f69bf2c48d85498b2f26e26361492896446a09bffa30a4abdfe278bbdf8c: Status 404 returned error can't find the container with id 3948f69bf2c48d85498b2f26e26361492896446a09bffa30a4abdfe278bbdf8c Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.460390 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" event={"ID":"032f4b3b-1037-4237-ba70-77d3f9753d9e","Type":"ContainerDied","Data":"332a83e0ad54ec6077318b3f87e89f82ef52db51a8f24b6f83a1a8fc8a540062"} Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.459564 4573 generic.go:334] "Generic (PLEG): container finished" podID="032f4b3b-1037-4237-ba70-77d3f9753d9e" containerID="332a83e0ad54ec6077318b3f87e89f82ef52db51a8f24b6f83a1a8fc8a540062" exitCode=0 Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.466476 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" event={"ID":"032f4b3b-1037-4237-ba70-77d3f9753d9e","Type":"ContainerStarted","Data":"763232cd5df7c66e0c15c403d05ff00a759e7b1c9e5b87e11457bb0d923a3560"} Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.572833 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66f99c579b-gh7sh" event={"ID":"b78e2453-adc3-4139-97c5-9f8a977df1ee","Type":"ContainerStarted","Data":"e4e59bc781e00123addd009f9b5eb99f124bfa5f54e8ec7fc50794233ac1821b"} Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.582255 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerStarted","Data":"39cb62ba4ad569a45049cde16bc1cf40005dce1875fe10337e5b5ae8bcb634a7"} Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.664516 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-fd87c"] Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.680356 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53de8a70-c359-4b86-a783-cf1815b73d68","Type":"ContainerStarted","Data":"c4c68bcf7251acc1034639d03065ae89fd34ee08fa1c5a732eac7d35b50cd68f"} Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.705224 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8016ac24-adaf-481c-8ad2-6016bc588fa9","Type":"ContainerStarted","Data":"3948f69bf2c48d85498b2f26e26361492896446a09bffa30a4abdfe278bbdf8c"} Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.777948 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-kvbg6" event={"ID":"df6a51aa-75ab-4bdc-9e22-6ba714772ecf","Type":"ContainerStarted","Data":"eb1fea2c39e1971637b2fb8d393068ecb9d4e67e5655a0afecace4db4cd54ca6"} Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.811759 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-v87zn"] Dec 03 09:00:53 crc kubenswrapper[4573]: E1203 09:00:53.812313 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8386e5-315f-4d6f-a560-e1a45777e339" containerName="neutron-db-sync" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.812331 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8386e5-315f-4d6f-a560-e1a45777e339" containerName="neutron-db-sync" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.812543 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8386e5-315f-4d6f-a560-e1a45777e339" containerName="neutron-db-sync" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.813541 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.832262 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m4vz4" event={"ID":"e28fb330-9e1a-4480-9a2d-b059d8e1576d","Type":"ContainerStarted","Data":"a831b5f7a5bb305ff88ecadd859e4cf44257739ca607a0df313d47dc032d6cc2"} Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.840977 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-v87zn"] Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.849240 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-kvbg6" podStartSLOduration=7.744461168 podStartE2EDuration="54.849202305s" podCreationTimestamp="2025-12-03 08:59:59 +0000 UTC" firstStartedPulling="2025-12-03 09:00:04.449063686 +0000 UTC m=+1325.017442935" lastFinishedPulling="2025-12-03 09:00:51.553804803 +0000 UTC m=+1372.122184072" observedRunningTime="2025-12-03 09:00:53.836973067 +0000 UTC m=+1374.405352326" watchObservedRunningTime="2025-12-03 09:00:53.849202305 +0000 UTC m=+1374.417581564" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.919990 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-m4vz4" podStartSLOduration=34.919951192 podStartE2EDuration="34.919951192s" podCreationTimestamp="2025-12-03 09:00:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:00:53.879451127 +0000 UTC m=+1374.447830386" watchObservedRunningTime="2025-12-03 09:00:53.919951192 +0000 UTC m=+1374.488330451" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.922827 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.922969 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-config\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.923135 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdftc\" (UniqueName: \"kubernetes.io/projected/317e24d7-2fe9-412b-8fb7-1ccb697b12de-kube-api-access-pdftc\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.923235 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-svc\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.923301 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.923342 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.967366 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-756bc5c75b-vdw4l"] Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.968938 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.974702 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.977672 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.978134 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 09:00:53 crc kubenswrapper[4573]: I1203 09:00:53.978345 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-255sb" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.022209 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-756bc5c75b-vdw4l"] Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040101 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040220 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-config\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040258 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-ovndb-tls-certs\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040374 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-httpd-config\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040399 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-config\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040473 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdftc\" (UniqueName: \"kubernetes.io/projected/317e24d7-2fe9-412b-8fb7-1ccb697b12de-kube-api-access-pdftc\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040536 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-combined-ca-bundle\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040584 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqlxt\" (UniqueName: \"kubernetes.io/projected/574fcf75-461a-4cfa-b1e0-d9759ee7be40-kube-api-access-nqlxt\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040605 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-svc\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040663 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.040722 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.041735 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.042319 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.042890 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-config\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.045929 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.046710 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-svc\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.113864 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e511768a-604b-441d-a363-e5c522b39647" path="/var/lib/kubelet/pods/e511768a-604b-441d-a363-e5c522b39647/volumes" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.118507 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdftc\" (UniqueName: \"kubernetes.io/projected/317e24d7-2fe9-412b-8fb7-1ccb697b12de-kube-api-access-pdftc\") pod \"dnsmasq-dns-6b7b667979-v87zn\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.142973 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-ovndb-tls-certs\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.143133 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-httpd-config\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.143165 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-config\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.143694 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-combined-ca-bundle\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.143752 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqlxt\" (UniqueName: \"kubernetes.io/projected/574fcf75-461a-4cfa-b1e0-d9759ee7be40-kube-api-access-nqlxt\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.157159 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.172890 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-ovndb-tls-certs\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.173463 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-config\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.185767 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-combined-ca-bundle\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.193896 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-httpd-config\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.201252 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqlxt\" (UniqueName: \"kubernetes.io/projected/574fcf75-461a-4cfa-b1e0-d9759ee7be40-kube-api-access-nqlxt\") pod \"neutron-756bc5c75b-vdw4l\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:54 crc kubenswrapper[4573]: I1203 09:00:54.305478 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:55 crc kubenswrapper[4573]: I1203 09:00:55.046580 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66f99c579b-gh7sh" event={"ID":"b78e2453-adc3-4139-97c5-9f8a977df1ee","Type":"ContainerStarted","Data":"afd0f349dd0be138a212ad8426fb19507ce3fcb56dbe6dbd953d89456cd91458"} Dec 03 09:00:55 crc kubenswrapper[4573]: I1203 09:00:55.161084 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66f99c579b-gh7sh" podStartSLOduration=5.468992994 podStartE2EDuration="45.161028022s" podCreationTimestamp="2025-12-03 09:00:10 +0000 UTC" firstStartedPulling="2025-12-03 09:00:12.399961719 +0000 UTC m=+1332.968340978" lastFinishedPulling="2025-12-03 09:00:52.091996747 +0000 UTC m=+1372.660376006" observedRunningTime="2025-12-03 09:00:55.147161411 +0000 UTC m=+1375.715540690" watchObservedRunningTime="2025-12-03 09:00:55.161028022 +0000 UTC m=+1375.729407281" Dec 03 09:00:55 crc kubenswrapper[4573]: I1203 09:00:55.208700 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-v87zn"] Dec 03 09:00:55 crc kubenswrapper[4573]: E1203 09:00:55.543184 4573 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 03 09:00:55 crc kubenswrapper[4573]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/032f4b3b-1037-4237-ba70-77d3f9753d9e/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 09:00:55 crc kubenswrapper[4573]: > podSandboxID="763232cd5df7c66e0c15c403d05ff00a759e7b1c9e5b87e11457bb0d923a3560" Dec 03 09:00:55 crc kubenswrapper[4573]: E1203 09:00:55.543924 4573 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 03 09:00:55 crc kubenswrapper[4573]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5dbh55fhfh596h68dh54dh55ch68ch5c7h5f8h68ch58fh644hb8h65bhc5h5d9h5bdhdh577h5dbh64dh694h687h667hcdh664h695h6fh579h95h565q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-njkwv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-56df8fb6b7-fd87c_openstack(032f4b3b-1037-4237-ba70-77d3f9753d9e): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/032f4b3b-1037-4237-ba70-77d3f9753d9e/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 09:00:55 crc kubenswrapper[4573]: > logger="UnhandledError" Dec 03 09:00:55 crc kubenswrapper[4573]: E1203 09:00:55.545182 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/032f4b3b-1037-4237-ba70-77d3f9753d9e/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" podUID="032f4b3b-1037-4237-ba70-77d3f9753d9e" Dec 03 09:00:56 crc kubenswrapper[4573]: I1203 09:00:56.015791 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-756bc5c75b-vdw4l"] Dec 03 09:00:56 crc kubenswrapper[4573]: W1203 09:00:56.054394 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod574fcf75_461a_4cfa_b1e0_d9759ee7be40.slice/crio-9f2c13ff514b3a75bc91efb9fbc98c8ab8b6501081e3b9794298cb1e4fdc4af3 WatchSource:0}: Error finding container 9f2c13ff514b3a75bc91efb9fbc98c8ab8b6501081e3b9794298cb1e4fdc4af3: Status 404 returned error can't find the container with id 9f2c13ff514b3a75bc91efb9fbc98c8ab8b6501081e3b9794298cb1e4fdc4af3 Dec 03 09:00:56 crc kubenswrapper[4573]: I1203 09:00:56.136758 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerStarted","Data":"875e1f86deecceeb9c350e1909663db37b4f6f0a82cff5c1165cad17d9cdc1df"} Dec 03 09:00:56 crc kubenswrapper[4573]: I1203 09:00:56.161868 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53de8a70-c359-4b86-a783-cf1815b73d68","Type":"ContainerStarted","Data":"bb8acd472ba37e6f329d6fdc15fb503b1e314810b71450a8a2a0d02841ecae76"} Dec 03 09:00:56 crc kubenswrapper[4573]: I1203 09:00:56.169465 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" event={"ID":"317e24d7-2fe9-412b-8fb7-1ccb697b12de","Type":"ContainerStarted","Data":"30e65a70ace00f0cd9c6f570526aaced5a203cc0a3d8dcb4a300b90b0d00b509"} Dec 03 09:00:56 crc kubenswrapper[4573]: I1203 09:00:56.179956 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b8b5b6556-4zxxs" podStartSLOduration=6.690534692 podStartE2EDuration="46.179925538s" podCreationTimestamp="2025-12-03 09:00:10 +0000 UTC" firstStartedPulling="2025-12-03 09:00:12.721987737 +0000 UTC m=+1333.290367006" lastFinishedPulling="2025-12-03 09:00:52.211378593 +0000 UTC m=+1372.779757852" observedRunningTime="2025-12-03 09:00:56.170305832 +0000 UTC m=+1376.738685091" watchObservedRunningTime="2025-12-03 09:00:56.179925538 +0000 UTC m=+1376.748304797" Dec 03 09:00:56 crc kubenswrapper[4573]: I1203 09:00:56.202214 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8016ac24-adaf-481c-8ad2-6016bc588fa9","Type":"ContainerStarted","Data":"911429dbe51a50ebfb10b3b0c0ce4572e1872dbb928363122b7de9fc1a6d05d2"} Dec 03 09:00:56 crc kubenswrapper[4573]: I1203 09:00:56.974361 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.123358 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-nb\") pod \"032f4b3b-1037-4237-ba70-77d3f9753d9e\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.123476 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-sb\") pod \"032f4b3b-1037-4237-ba70-77d3f9753d9e\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.123708 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njkwv\" (UniqueName: \"kubernetes.io/projected/032f4b3b-1037-4237-ba70-77d3f9753d9e-kube-api-access-njkwv\") pod \"032f4b3b-1037-4237-ba70-77d3f9753d9e\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.123744 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-svc\") pod \"032f4b3b-1037-4237-ba70-77d3f9753d9e\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.123820 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-config\") pod \"032f4b3b-1037-4237-ba70-77d3f9753d9e\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.123898 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-swift-storage-0\") pod \"032f4b3b-1037-4237-ba70-77d3f9753d9e\" (UID: \"032f4b3b-1037-4237-ba70-77d3f9753d9e\") " Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.245943 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032f4b3b-1037-4237-ba70-77d3f9753d9e-kube-api-access-njkwv" (OuterVolumeSpecName: "kube-api-access-njkwv") pod "032f4b3b-1037-4237-ba70-77d3f9753d9e" (UID: "032f4b3b-1037-4237-ba70-77d3f9753d9e"). InnerVolumeSpecName "kube-api-access-njkwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.289496 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53de8a70-c359-4b86-a783-cf1815b73d68","Type":"ContainerStarted","Data":"7856c8b1de124b63cac54fd5bac549c4ccecc7895e45e53e2d0c6d7f978dd57b"} Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.291382 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" containerName="glance-log" containerID="cri-o://bb8acd472ba37e6f329d6fdc15fb503b1e314810b71450a8a2a0d02841ecae76" gracePeriod=30 Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.292939 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" containerName="glance-httpd" containerID="cri-o://7856c8b1de124b63cac54fd5bac549c4ccecc7895e45e53e2d0c6d7f978dd57b" gracePeriod=30 Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.308583 4573 generic.go:334] "Generic (PLEG): container finished" podID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" containerID="5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629" exitCode=0 Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.308728 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" event={"ID":"317e24d7-2fe9-412b-8fb7-1ccb697b12de","Type":"ContainerDied","Data":"5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629"} Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.319662 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerStarted","Data":"cb8dd971f575f2e4c8253374b56b43818da2c0e0bb78fb0073202f4cd1f5eb47"} Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.319958 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerStarted","Data":"9f2c13ff514b3a75bc91efb9fbc98c8ab8b6501081e3b9794298cb1e4fdc4af3"} Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.334983 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=25.334949169 podStartE2EDuration="25.334949169s" podCreationTimestamp="2025-12-03 09:00:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:00:57.321745646 +0000 UTC m=+1377.890124915" watchObservedRunningTime="2025-12-03 09:00:57.334949169 +0000 UTC m=+1377.903328428" Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.343278 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njkwv\" (UniqueName: \"kubernetes.io/projected/032f4b3b-1037-4237-ba70-77d3f9753d9e-kube-api-access-njkwv\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.347029 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.347272 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-fd87c" event={"ID":"032f4b3b-1037-4237-ba70-77d3f9753d9e","Type":"ContainerDied","Data":"763232cd5df7c66e0c15c403d05ff00a759e7b1c9e5b87e11457bb0d923a3560"} Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.347329 4573 scope.go:117] "RemoveContainer" containerID="332a83e0ad54ec6077318b3f87e89f82ef52db51a8f24b6f83a1a8fc8a540062" Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.715367 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "032f4b3b-1037-4237-ba70-77d3f9753d9e" (UID: "032f4b3b-1037-4237-ba70-77d3f9753d9e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:57 crc kubenswrapper[4573]: I1203 09:00:57.752444 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.157640 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "032f4b3b-1037-4237-ba70-77d3f9753d9e" (UID: "032f4b3b-1037-4237-ba70-77d3f9753d9e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.169179 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.337675 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-config" (OuterVolumeSpecName: "config") pod "032f4b3b-1037-4237-ba70-77d3f9753d9e" (UID: "032f4b3b-1037-4237-ba70-77d3f9753d9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.376827 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.418170 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "032f4b3b-1037-4237-ba70-77d3f9753d9e" (UID: "032f4b3b-1037-4237-ba70-77d3f9753d9e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.427760 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "032f4b3b-1037-4237-ba70-77d3f9753d9e" (UID: "032f4b3b-1037-4237-ba70-77d3f9753d9e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.450503 4573 generic.go:334] "Generic (PLEG): container finished" podID="53de8a70-c359-4b86-a783-cf1815b73d68" containerID="7856c8b1de124b63cac54fd5bac549c4ccecc7895e45e53e2d0c6d7f978dd57b" exitCode=143 Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.450552 4573 generic.go:334] "Generic (PLEG): container finished" podID="53de8a70-c359-4b86-a783-cf1815b73d68" containerID="bb8acd472ba37e6f329d6fdc15fb503b1e314810b71450a8a2a0d02841ecae76" exitCode=143 Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.481958 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-756bc5c75b-vdw4l" podStartSLOduration=5.4819093389999995 podStartE2EDuration="5.481909339s" podCreationTimestamp="2025-12-03 09:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:00:58.480678296 +0000 UTC m=+1379.049057555" watchObservedRunningTime="2025-12-03 09:00:58.481909339 +0000 UTC m=+1379.050288598" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.482212 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.482251 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/032f4b3b-1037-4237-ba70-77d3f9753d9e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.583189 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.583243 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cfccd4555-2shp6"] Dec 03 09:00:58 crc kubenswrapper[4573]: E1203 09:00:58.583773 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032f4b3b-1037-4237-ba70-77d3f9753d9e" containerName="init" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.583790 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="032f4b3b-1037-4237-ba70-77d3f9753d9e" containerName="init" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.584064 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="032f4b3b-1037-4237-ba70-77d3f9753d9e" containerName="init" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.589271 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cfccd4555-2shp6"] Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.589311 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerStarted","Data":"d7432509950ad4b8f82aaee8374a0875e60a46c346d901ce1377748060c37945"} Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.589342 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53de8a70-c359-4b86-a783-cf1815b73d68","Type":"ContainerDied","Data":"7856c8b1de124b63cac54fd5bac549c4ccecc7895e45e53e2d0c6d7f978dd57b"} Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.589366 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53de8a70-c359-4b86-a783-cf1815b73d68","Type":"ContainerDied","Data":"bb8acd472ba37e6f329d6fdc15fb503b1e314810b71450a8a2a0d02841ecae76"} Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.589502 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.600633 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.602314 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.616716 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-combined-ca-bundle\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.616791 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-httpd-config\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.616872 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-ovndb-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.616984 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-public-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.617084 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-config\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.617240 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-internal-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.617282 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmr9j\" (UniqueName: \"kubernetes.io/projected/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-kube-api-access-mmr9j\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.719322 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-combined-ca-bundle\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.719903 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-httpd-config\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.719946 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-ovndb-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.720015 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-public-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.728335 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-config\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.728627 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-internal-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.728666 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmr9j\" (UniqueName: \"kubernetes.io/projected/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-kube-api-access-mmr9j\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.774506 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmr9j\" (UniqueName: \"kubernetes.io/projected/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-kube-api-access-mmr9j\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.784409 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-fd87c"] Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.811023 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-internal-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.811029 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-combined-ca-bundle\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.811579 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-httpd-config\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.813033 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-config\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.816440 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-ovndb-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.820679 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4434273d-ce2e-4ecd-820d-7f2ecf35bca7-public-tls-certs\") pod \"neutron-6cfccd4555-2shp6\" (UID: \"4434273d-ce2e-4ecd-820d-7f2ecf35bca7\") " pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.832611 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-fd87c"] Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.952529 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:00:58 crc kubenswrapper[4573]: I1203 09:00:58.967407 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.213107 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch7ws\" (UniqueName: \"kubernetes.io/projected/53de8a70-c359-4b86-a783-cf1815b73d68-kube-api-access-ch7ws\") pod \"53de8a70-c359-4b86-a783-cf1815b73d68\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.213592 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-config-data\") pod \"53de8a70-c359-4b86-a783-cf1815b73d68\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.213653 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-scripts\") pod \"53de8a70-c359-4b86-a783-cf1815b73d68\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.213694 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-httpd-run\") pod \"53de8a70-c359-4b86-a783-cf1815b73d68\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.213730 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-combined-ca-bundle\") pod \"53de8a70-c359-4b86-a783-cf1815b73d68\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.213798 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-logs\") pod \"53de8a70-c359-4b86-a783-cf1815b73d68\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.213824 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"53de8a70-c359-4b86-a783-cf1815b73d68\" (UID: \"53de8a70-c359-4b86-a783-cf1815b73d68\") " Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.214792 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-logs" (OuterVolumeSpecName: "logs") pod "53de8a70-c359-4b86-a783-cf1815b73d68" (UID: "53de8a70-c359-4b86-a783-cf1815b73d68"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.217288 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "53de8a70-c359-4b86-a783-cf1815b73d68" (UID: "53de8a70-c359-4b86-a783-cf1815b73d68"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.267244 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "53de8a70-c359-4b86-a783-cf1815b73d68" (UID: "53de8a70-c359-4b86-a783-cf1815b73d68"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.269614 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-scripts" (OuterVolumeSpecName: "scripts") pod "53de8a70-c359-4b86-a783-cf1815b73d68" (UID: "53de8a70-c359-4b86-a783-cf1815b73d68"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.310723 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53de8a70-c359-4b86-a783-cf1815b73d68-kube-api-access-ch7ws" (OuterVolumeSpecName: "kube-api-access-ch7ws") pod "53de8a70-c359-4b86-a783-cf1815b73d68" (UID: "53de8a70-c359-4b86-a783-cf1815b73d68"). InnerVolumeSpecName "kube-api-access-ch7ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.316379 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.317579 4573 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.317660 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch7ws\" (UniqueName: \"kubernetes.io/projected/53de8a70-c359-4b86-a783-cf1815b73d68-kube-api-access-ch7ws\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.317756 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.317831 4573 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53de8a70-c359-4b86-a783-cf1815b73d68-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.510487 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53de8a70-c359-4b86-a783-cf1815b73d68" (UID: "53de8a70-c359-4b86-a783-cf1815b73d68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.524079 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.546790 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53de8a70-c359-4b86-a783-cf1815b73d68","Type":"ContainerDied","Data":"c4c68bcf7251acc1034639d03065ae89fd34ee08fa1c5a732eac7d35b50cd68f"} Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.547135 4573 scope.go:117] "RemoveContainer" containerID="7856c8b1de124b63cac54fd5bac549c4ccecc7895e45e53e2d0c6d7f978dd57b" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.547323 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.557365 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-config-data" (OuterVolumeSpecName: "config-data") pod "53de8a70-c359-4b86-a783-cf1815b73d68" (UID: "53de8a70-c359-4b86-a783-cf1815b73d68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.560778 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerName="glance-log" containerID="cri-o://911429dbe51a50ebfb10b3b0c0ce4572e1872dbb928363122b7de9fc1a6d05d2" gracePeriod=30 Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.563177 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8016ac24-adaf-481c-8ad2-6016bc588fa9","Type":"ContainerStarted","Data":"9e1215b1bf5d058552a1f2196ba73fb04c702ca99d7c1cc61379d614f401b27c"} Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.563262 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerName="glance-httpd" containerID="cri-o://9e1215b1bf5d058552a1f2196ba73fb04c702ca99d7c1cc61379d614f401b27c" gracePeriod=30 Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.635364 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53de8a70-c359-4b86-a783-cf1815b73d68-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.650181 4573 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.722793 4573 scope.go:117] "RemoveContainer" containerID="bb8acd472ba37e6f329d6fdc15fb503b1e314810b71450a8a2a0d02841ecae76" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.742294 4573 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.955258 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.955224432 podStartE2EDuration="27.955224432s" podCreationTimestamp="2025-12-03 09:00:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:00:59.633725164 +0000 UTC m=+1380.202104423" watchObservedRunningTime="2025-12-03 09:00:59.955224432 +0000 UTC m=+1380.523603691" Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.970310 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:00:59 crc kubenswrapper[4573]: I1203 09:00:59.980527 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.028086 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cfccd4555-2shp6"] Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.056638 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="032f4b3b-1037-4237-ba70-77d3f9753d9e" path="/var/lib/kubelet/pods/032f4b3b-1037-4237-ba70-77d3f9753d9e/volumes" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.057646 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" path="/var/lib/kubelet/pods/53de8a70-c359-4b86-a783-cf1815b73d68/volumes" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.146844 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:01:00 crc kubenswrapper[4573]: E1203 09:01:00.152247 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" containerName="glance-log" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.152285 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" containerName="glance-log" Dec 03 09:01:00 crc kubenswrapper[4573]: E1203 09:01:00.152326 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" containerName="glance-httpd" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.152341 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" containerName="glance-httpd" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.159663 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" containerName="glance-httpd" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.159725 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="53de8a70-c359-4b86-a783-cf1815b73d68" containerName="glance-log" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.177852 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.210044 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.210504 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.218250 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.280942 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.281023 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.281075 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-logs\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.282430 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.282545 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.282619 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.282723 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm2dl\" (UniqueName: \"kubernetes.io/projected/50863c47-12be-414c-9bd9-fdb5347fea7a-kube-api-access-sm2dl\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.283178 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.387071 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.387174 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.387237 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.387286 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm2dl\" (UniqueName: \"kubernetes.io/projected/50863c47-12be-414c-9bd9-fdb5347fea7a-kube-api-access-sm2dl\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.387350 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.387411 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.387438 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.387463 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-logs\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.388108 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-logs\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.393914 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.394538 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.400258 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.406731 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.409321 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.409376 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.419990 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm2dl\" (UniqueName: \"kubernetes.io/projected/50863c47-12be-414c-9bd9-fdb5347fea7a-kube-api-access-sm2dl\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.464164 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.586928 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cfccd4555-2shp6" event={"ID":"4434273d-ce2e-4ecd-820d-7f2ecf35bca7","Type":"ContainerStarted","Data":"6ddcfc61b0a5526b0b96a3f93a82629cbc14a22ef5b3f321efd764819b44fd76"} Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.598315 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" event={"ID":"317e24d7-2fe9-412b-8fb7-1ccb697b12de","Type":"ContainerStarted","Data":"71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd"} Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.598922 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.618632 4573 generic.go:334] "Generic (PLEG): container finished" podID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerID="9e1215b1bf5d058552a1f2196ba73fb04c702ca99d7c1cc61379d614f401b27c" exitCode=143 Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.618708 4573 generic.go:334] "Generic (PLEG): container finished" podID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerID="911429dbe51a50ebfb10b3b0c0ce4572e1872dbb928363122b7de9fc1a6d05d2" exitCode=143 Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.618732 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8016ac24-adaf-481c-8ad2-6016bc588fa9","Type":"ContainerDied","Data":"9e1215b1bf5d058552a1f2196ba73fb04c702ca99d7c1cc61379d614f401b27c"} Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.618825 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8016ac24-adaf-481c-8ad2-6016bc588fa9","Type":"ContainerDied","Data":"911429dbe51a50ebfb10b3b0c0ce4572e1872dbb928363122b7de9fc1a6d05d2"} Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.623631 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/0.log" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.625851 4573 generic.go:334] "Generic (PLEG): container finished" podID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerID="d7432509950ad4b8f82aaee8374a0875e60a46c346d901ce1377748060c37945" exitCode=1 Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.625923 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerDied","Data":"d7432509950ad4b8f82aaee8374a0875e60a46c346d901ce1377748060c37945"} Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.627113 4573 scope.go:117] "RemoveContainer" containerID="d7432509950ad4b8f82aaee8374a0875e60a46c346d901ce1377748060c37945" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.663109 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" podStartSLOduration=7.6630738560000005 podStartE2EDuration="7.663073856s" podCreationTimestamp="2025-12-03 09:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:00.640394732 +0000 UTC m=+1381.208773991" watchObservedRunningTime="2025-12-03 09:01:00.663073856 +0000 UTC m=+1381.231453115" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.669533 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.918538 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:01:00 crc kubenswrapper[4573]: I1203 09:01:00.919186 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.071966 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.072035 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.377831 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.425421 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-config-data\") pod \"8016ac24-adaf-481c-8ad2-6016bc588fa9\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.425591 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gf6m\" (UniqueName: \"kubernetes.io/projected/8016ac24-adaf-481c-8ad2-6016bc588fa9-kube-api-access-5gf6m\") pod \"8016ac24-adaf-481c-8ad2-6016bc588fa9\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.425832 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-logs\") pod \"8016ac24-adaf-481c-8ad2-6016bc588fa9\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.425920 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-combined-ca-bundle\") pod \"8016ac24-adaf-481c-8ad2-6016bc588fa9\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.425963 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-httpd-run\") pod \"8016ac24-adaf-481c-8ad2-6016bc588fa9\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.426062 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"8016ac24-adaf-481c-8ad2-6016bc588fa9\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.426081 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-scripts\") pod \"8016ac24-adaf-481c-8ad2-6016bc588fa9\" (UID: \"8016ac24-adaf-481c-8ad2-6016bc588fa9\") " Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.427914 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-logs" (OuterVolumeSpecName: "logs") pod "8016ac24-adaf-481c-8ad2-6016bc588fa9" (UID: "8016ac24-adaf-481c-8ad2-6016bc588fa9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.429480 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8016ac24-adaf-481c-8ad2-6016bc588fa9" (UID: "8016ac24-adaf-481c-8ad2-6016bc588fa9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.434532 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.434597 4573 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8016ac24-adaf-481c-8ad2-6016bc588fa9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.461015 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-scripts" (OuterVolumeSpecName: "scripts") pod "8016ac24-adaf-481c-8ad2-6016bc588fa9" (UID: "8016ac24-adaf-481c-8ad2-6016bc588fa9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.461171 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "8016ac24-adaf-481c-8ad2-6016bc588fa9" (UID: "8016ac24-adaf-481c-8ad2-6016bc588fa9"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.471320 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8016ac24-adaf-481c-8ad2-6016bc588fa9-kube-api-access-5gf6m" (OuterVolumeSpecName: "kube-api-access-5gf6m") pod "8016ac24-adaf-481c-8ad2-6016bc588fa9" (UID: "8016ac24-adaf-481c-8ad2-6016bc588fa9"). InnerVolumeSpecName "kube-api-access-5gf6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.536503 4573 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.537019 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.537033 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gf6m\" (UniqueName: \"kubernetes.io/projected/8016ac24-adaf-481c-8ad2-6016bc588fa9-kube-api-access-5gf6m\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.748196 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/0.log" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.764515 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerStarted","Data":"c356ec893e6d00cfbf39cf9f08708310b04998f0f0aa8654a1b0daa62c4db295"} Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.766618 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.771202 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8016ac24-adaf-481c-8ad2-6016bc588fa9" (UID: "8016ac24-adaf-481c-8ad2-6016bc588fa9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.793327 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cfccd4555-2shp6" event={"ID":"4434273d-ce2e-4ecd-820d-7f2ecf35bca7","Type":"ContainerStarted","Data":"ddf2e5351cabc58a8084625aa1d6d8a8dd076c3bf3da034ecfaa0c2a065a56d2"} Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.828537 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-config-data" (OuterVolumeSpecName: "config-data") pod "8016ac24-adaf-481c-8ad2-6016bc588fa9" (UID: "8016ac24-adaf-481c-8ad2-6016bc588fa9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.836314 4573 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.844291 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.845258 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8016ac24-adaf-481c-8ad2-6016bc588fa9","Type":"ContainerDied","Data":"3948f69bf2c48d85498b2f26e26361492896446a09bffa30a4abdfe278bbdf8c"} Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.845525 4573 scope.go:117] "RemoveContainer" containerID="9e1215b1bf5d058552a1f2196ba73fb04c702ca99d7c1cc61379d614f401b27c" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.859823 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.859869 4573 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.859879 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8016ac24-adaf-481c-8ad2-6016bc588fa9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.948150 4573 scope.go:117] "RemoveContainer" containerID="911429dbe51a50ebfb10b3b0c0ce4572e1872dbb928363122b7de9fc1a6d05d2" Dec 03 09:01:01 crc kubenswrapper[4573]: I1203 09:01:01.977912 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.000780 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.055989 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" path="/var/lib/kubelet/pods/8016ac24-adaf-481c-8ad2-6016bc588fa9/volumes" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.056784 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:01:02 crc kubenswrapper[4573]: E1203 09:01:02.057731 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerName="glance-log" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.057766 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerName="glance-log" Dec 03 09:01:02 crc kubenswrapper[4573]: E1203 09:01:02.057836 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerName="glance-httpd" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.057852 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerName="glance-httpd" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.062852 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerName="glance-httpd" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.062961 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8016ac24-adaf-481c-8ad2-6016bc588fa9" containerName="glance-log" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.064618 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.081779 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.081785 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.089502 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.104158 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:01:02 crc kubenswrapper[4573]: W1203 09:01:02.130627 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50863c47_12be_414c_9bd9_fdb5347fea7a.slice/crio-59610d98baec5796929f6e447112192b2af6c68199d48d2762725f6ac0d0773e WatchSource:0}: Error finding container 59610d98baec5796929f6e447112192b2af6c68199d48d2762725f6ac0d0773e: Status 404 returned error can't find the container with id 59610d98baec5796929f6e447112192b2af6c68199d48d2762725f6ac0d0773e Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.168550 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.168602 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.168636 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.168672 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-logs\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.168722 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.168778 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.168798 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.168833 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkv6s\" (UniqueName: \"kubernetes.io/projected/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-kube-api-access-mkv6s\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.272813 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.272879 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.272910 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.272947 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-logs\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.272995 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.273034 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.273099 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.273146 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkv6s\" (UniqueName: \"kubernetes.io/projected/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-kube-api-access-mkv6s\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.278672 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.518039 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.518100 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-logs\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.531230 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkv6s\" (UniqueName: \"kubernetes.io/projected/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-kube-api-access-mkv6s\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.531959 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.555992 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.566869 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.574824 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.714898 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.785463 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.958134 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50863c47-12be-414c-9bd9-fdb5347fea7a","Type":"ContainerStarted","Data":"59610d98baec5796929f6e447112192b2af6c68199d48d2762725f6ac0d0773e"} Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.986367 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cfccd4555-2shp6" event={"ID":"4434273d-ce2e-4ecd-820d-7f2ecf35bca7","Type":"ContainerStarted","Data":"469de7d3bb23752de2fbae3b7f1844e10a20b9d5e0701015f826379e120cf63f"} Dec 03 09:01:02 crc kubenswrapper[4573]: I1203 09:01:02.986432 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:01:03 crc kubenswrapper[4573]: I1203 09:01:03.027453 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cfccd4555-2shp6" podStartSLOduration=5.027399654 podStartE2EDuration="5.027399654s" podCreationTimestamp="2025-12-03 09:00:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:03.014369555 +0000 UTC m=+1383.582748824" watchObservedRunningTime="2025-12-03 09:01:03.027399654 +0000 UTC m=+1383.595778913" Dec 03 09:01:03 crc kubenswrapper[4573]: I1203 09:01:03.911297 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:01:03 crc kubenswrapper[4573]: W1203 09:01:03.925178 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f33c171_b5c5_444e_9cd0_63e2a2322f1c.slice/crio-196a3fd162473cb4c1ddfb831ba268afae2b1bb7f0c12d97a1fbd7af61aae0a1 WatchSource:0}: Error finding container 196a3fd162473cb4c1ddfb831ba268afae2b1bb7f0c12d97a1fbd7af61aae0a1: Status 404 returned error can't find the container with id 196a3fd162473cb4c1ddfb831ba268afae2b1bb7f0c12d97a1fbd7af61aae0a1 Dec 03 09:01:03 crc kubenswrapper[4573]: I1203 09:01:03.999160 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/1.log" Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.002299 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/0.log" Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.003224 4573 generic.go:334] "Generic (PLEG): container finished" podID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerID="c356ec893e6d00cfbf39cf9f08708310b04998f0f0aa8654a1b0daa62c4db295" exitCode=1 Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.003697 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerDied","Data":"c356ec893e6d00cfbf39cf9f08708310b04998f0f0aa8654a1b0daa62c4db295"} Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.003749 4573 scope.go:117] "RemoveContainer" containerID="d7432509950ad4b8f82aaee8374a0875e60a46c346d901ce1377748060c37945" Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.005105 4573 scope.go:117] "RemoveContainer" containerID="c356ec893e6d00cfbf39cf9f08708310b04998f0f0aa8654a1b0daa62c4db295" Dec 03 09:01:04 crc kubenswrapper[4573]: E1203 09:01:04.005402 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 10s restarting failed container=neutron-httpd pod=neutron-756bc5c75b-vdw4l_openstack(574fcf75-461a-4cfa-b1e0-d9759ee7be40)\"" pod="openstack/neutron-756bc5c75b-vdw4l" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.010111 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8f33c171-b5c5-444e-9cd0-63e2a2322f1c","Type":"ContainerStarted","Data":"196a3fd162473cb4c1ddfb831ba268afae2b1bb7f0c12d97a1fbd7af61aae0a1"} Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.022727 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50863c47-12be-414c-9bd9-fdb5347fea7a","Type":"ContainerStarted","Data":"0d7f332fa30855ffec74354857b86688208d19f2199b44d2ebb3d88b00a73383"} Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.164572 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.275304 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-jckvv"] Dec 03 09:01:04 crc kubenswrapper[4573]: I1203 09:01:04.275793 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerName="dnsmasq-dns" containerID="cri-o://fbf6aa1c5cec0d6236795a987b1189668404b7ceae8aa0b0a0cfbdc073a0aaf7" gracePeriod=10 Dec 03 09:01:05 crc kubenswrapper[4573]: I1203 09:01:05.070567 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" event={"ID":"1f1b04c5-a752-4557-a537-ad477d6d6280","Type":"ContainerDied","Data":"fbf6aa1c5cec0d6236795a987b1189668404b7ceae8aa0b0a0cfbdc073a0aaf7"} Dec 03 09:01:05 crc kubenswrapper[4573]: I1203 09:01:05.070371 4573 generic.go:334] "Generic (PLEG): container finished" podID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerID="fbf6aa1c5cec0d6236795a987b1189668404b7ceae8aa0b0a0cfbdc073a0aaf7" exitCode=0 Dec 03 09:01:05 crc kubenswrapper[4573]: I1203 09:01:05.075662 4573 generic.go:334] "Generic (PLEG): container finished" podID="df6a51aa-75ab-4bdc-9e22-6ba714772ecf" containerID="eb1fea2c39e1971637b2fb8d393068ecb9d4e67e5655a0afecace4db4cd54ca6" exitCode=0 Dec 03 09:01:05 crc kubenswrapper[4573]: I1203 09:01:05.075749 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-kvbg6" event={"ID":"df6a51aa-75ab-4bdc-9e22-6ba714772ecf","Type":"ContainerDied","Data":"eb1fea2c39e1971637b2fb8d393068ecb9d4e67e5655a0afecace4db4cd54ca6"} Dec 03 09:01:05 crc kubenswrapper[4573]: I1203 09:01:05.082303 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/1.log" Dec 03 09:01:10 crc kubenswrapper[4573]: I1203 09:01:10.331786 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8f33c171-b5c5-444e-9cd0-63e2a2322f1c","Type":"ContainerStarted","Data":"1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287"} Dec 03 09:01:10 crc kubenswrapper[4573]: I1203 09:01:10.931914 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.035287 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: i/o timeout" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.074469 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.348264 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-kvbg6" event={"ID":"df6a51aa-75ab-4bdc-9e22-6ba714772ecf","Type":"ContainerDied","Data":"55a8f18aea332a924ed619229167131088fb4f1445db489d59c16f8e04990e0f"} Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.348324 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55a8f18aea332a924ed619229167131088fb4f1445db489d59c16f8e04990e0f" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.492771 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-kvbg6" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.503773 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645019 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9krcj\" (UniqueName: \"kubernetes.io/projected/1f1b04c5-a752-4557-a537-ad477d6d6280-kube-api-access-9krcj\") pod \"1f1b04c5-a752-4557-a537-ad477d6d6280\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645159 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-swift-storage-0\") pod \"1f1b04c5-a752-4557-a537-ad477d6d6280\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645197 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-sb\") pod \"1f1b04c5-a752-4557-a537-ad477d6d6280\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645321 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-config-data\") pod \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645359 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-logs\") pod \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645491 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-scripts\") pod \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645531 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-config\") pod \"1f1b04c5-a752-4557-a537-ad477d6d6280\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645599 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hhsj\" (UniqueName: \"kubernetes.io/projected/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-kube-api-access-4hhsj\") pod \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645635 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-combined-ca-bundle\") pod \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\" (UID: \"df6a51aa-75ab-4bdc-9e22-6ba714772ecf\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645667 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-svc\") pod \"1f1b04c5-a752-4557-a537-ad477d6d6280\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.645690 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-nb\") pod \"1f1b04c5-a752-4557-a537-ad477d6d6280\" (UID: \"1f1b04c5-a752-4557-a537-ad477d6d6280\") " Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.647626 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-logs" (OuterVolumeSpecName: "logs") pod "df6a51aa-75ab-4bdc-9e22-6ba714772ecf" (UID: "df6a51aa-75ab-4bdc-9e22-6ba714772ecf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.653881 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-scripts" (OuterVolumeSpecName: "scripts") pod "df6a51aa-75ab-4bdc-9e22-6ba714772ecf" (UID: "df6a51aa-75ab-4bdc-9e22-6ba714772ecf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.654459 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f1b04c5-a752-4557-a537-ad477d6d6280-kube-api-access-9krcj" (OuterVolumeSpecName: "kube-api-access-9krcj") pod "1f1b04c5-a752-4557-a537-ad477d6d6280" (UID: "1f1b04c5-a752-4557-a537-ad477d6d6280"). InnerVolumeSpecName "kube-api-access-9krcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.702196 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-kube-api-access-4hhsj" (OuterVolumeSpecName: "kube-api-access-4hhsj") pod "df6a51aa-75ab-4bdc-9e22-6ba714772ecf" (UID: "df6a51aa-75ab-4bdc-9e22-6ba714772ecf"). InnerVolumeSpecName "kube-api-access-4hhsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.718522 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df6a51aa-75ab-4bdc-9e22-6ba714772ecf" (UID: "df6a51aa-75ab-4bdc-9e22-6ba714772ecf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.747411 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-config-data" (OuterVolumeSpecName: "config-data") pod "df6a51aa-75ab-4bdc-9e22-6ba714772ecf" (UID: "df6a51aa-75ab-4bdc-9e22-6ba714772ecf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.748518 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hhsj\" (UniqueName: \"kubernetes.io/projected/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-kube-api-access-4hhsj\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.748598 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.748617 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9krcj\" (UniqueName: \"kubernetes.io/projected/1f1b04c5-a752-4557-a537-ad477d6d6280-kube-api-access-9krcj\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.748633 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.748644 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.748657 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df6a51aa-75ab-4bdc-9e22-6ba714772ecf-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.771820 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1f1b04c5-a752-4557-a537-ad477d6d6280" (UID: "1f1b04c5-a752-4557-a537-ad477d6d6280"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.779880 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1f1b04c5-a752-4557-a537-ad477d6d6280" (UID: "1f1b04c5-a752-4557-a537-ad477d6d6280"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.819775 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1f1b04c5-a752-4557-a537-ad477d6d6280" (UID: "1f1b04c5-a752-4557-a537-ad477d6d6280"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.821613 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1f1b04c5-a752-4557-a537-ad477d6d6280" (UID: "1f1b04c5-a752-4557-a537-ad477d6d6280"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.838087 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-config" (OuterVolumeSpecName: "config") pod "1f1b04c5-a752-4557-a537-ad477d6d6280" (UID: "1f1b04c5-a752-4557-a537-ad477d6d6280"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.855725 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.855771 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.855845 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.855858 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:11 crc kubenswrapper[4573]: I1203 09:01:11.855871 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f1b04c5-a752-4557-a537-ad477d6d6280-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.380041 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" event={"ID":"1f1b04c5-a752-4557-a537-ad477d6d6280","Type":"ContainerDied","Data":"a469fd84d82c8cd1438a1477c1d68221fb360047a50a0e0bb04e597e32b6a7c3"} Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.380138 4573 scope.go:117] "RemoveContainer" containerID="fbf6aa1c5cec0d6236795a987b1189668404b7ceae8aa0b0a0cfbdc073a0aaf7" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.381858 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.383039 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50863c47-12be-414c-9bd9-fdb5347fea7a","Type":"ContainerStarted","Data":"7f167bafe9560fae3cbca25f2f4f46eefa44a83cac71cba23ba0904e8c68bf7e"} Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.391670 4573 generic.go:334] "Generic (PLEG): container finished" podID="e28fb330-9e1a-4480-9a2d-b059d8e1576d" containerID="a831b5f7a5bb305ff88ecadd859e4cf44257739ca607a0df313d47dc032d6cc2" exitCode=0 Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.391758 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-kvbg6" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.392551 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m4vz4" event={"ID":"e28fb330-9e1a-4480-9a2d-b059d8e1576d","Type":"ContainerDied","Data":"a831b5f7a5bb305ff88ecadd859e4cf44257739ca607a0df313d47dc032d6cc2"} Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.450815 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=13.450776402 podStartE2EDuration="13.450776402s" podCreationTimestamp="2025-12-03 09:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:12.413595629 +0000 UTC m=+1392.981974898" watchObservedRunningTime="2025-12-03 09:01:12.450776402 +0000 UTC m=+1393.019155661" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.519108 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-jckvv"] Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.560135 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-jckvv"] Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.653675 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-85bbcf664-5fm4n"] Dec 03 09:01:12 crc kubenswrapper[4573]: E1203 09:01:12.654730 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerName="init" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.654754 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerName="init" Dec 03 09:01:12 crc kubenswrapper[4573]: E1203 09:01:12.654768 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df6a51aa-75ab-4bdc-9e22-6ba714772ecf" containerName="placement-db-sync" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.654774 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="df6a51aa-75ab-4bdc-9e22-6ba714772ecf" containerName="placement-db-sync" Dec 03 09:01:12 crc kubenswrapper[4573]: E1203 09:01:12.654783 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerName="dnsmasq-dns" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.654791 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerName="dnsmasq-dns" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.655039 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="df6a51aa-75ab-4bdc-9e22-6ba714772ecf" containerName="placement-db-sync" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.655085 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerName="dnsmasq-dns" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.656087 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.661676 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vtbrw" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.661785 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.661890 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.661993 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.662119 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.673003 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85bbcf664-5fm4n"] Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.715216 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-combined-ca-bundle\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.715278 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-public-tls-certs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.715309 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-internal-tls-certs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.715333 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-scripts\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.715400 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-config-data\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.715453 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b292e81d-5ded-4fff-bb9e-c764c395609b-logs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.715487 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znzlx\" (UniqueName: \"kubernetes.io/projected/b292e81d-5ded-4fff-bb9e-c764c395609b-kube-api-access-znzlx\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.817603 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-config-data\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.817681 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b292e81d-5ded-4fff-bb9e-c764c395609b-logs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.817721 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znzlx\" (UniqueName: \"kubernetes.io/projected/b292e81d-5ded-4fff-bb9e-c764c395609b-kube-api-access-znzlx\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.817789 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-combined-ca-bundle\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.817825 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-public-tls-certs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.817851 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-internal-tls-certs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.817876 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-scripts\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.818969 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b292e81d-5ded-4fff-bb9e-c764c395609b-logs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.825206 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-combined-ca-bundle\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.826250 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-public-tls-certs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.830911 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-scripts\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.833756 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-config-data\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.844896 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znzlx\" (UniqueName: \"kubernetes.io/projected/b292e81d-5ded-4fff-bb9e-c764c395609b-kube-api-access-znzlx\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:12 crc kubenswrapper[4573]: I1203 09:01:12.846585 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b292e81d-5ded-4fff-bb9e-c764c395609b-internal-tls-certs\") pod \"placement-85bbcf664-5fm4n\" (UID: \"b292e81d-5ded-4fff-bb9e-c764c395609b\") " pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:13 crc kubenswrapper[4573]: I1203 09:01:13.063316 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:14 crc kubenswrapper[4573]: I1203 09:01:14.046498 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" path="/var/lib/kubelet/pods/1f1b04c5-a752-4557-a537-ad477d6d6280/volumes" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.031644 4573 scope.go:117] "RemoveContainer" containerID="c356ec893e6d00cfbf39cf9f08708310b04998f0f0aa8654a1b0daa62c4db295" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.036007 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cf78879c9-jckvv" podUID="1f1b04c5-a752-4557-a537-ad477d6d6280" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: i/o timeout" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.117777 4573 scope.go:117] "RemoveContainer" containerID="47acf997deb639b2ea9e482a5ebdaa726ded6fcd513c9b477eb612e983e3de27" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.386649 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.504825 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-fernet-keys\") pod \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.504900 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-scripts\") pod \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.504972 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-credential-keys\") pod \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.505005 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-config-data\") pod \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.505080 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpnxr\" (UniqueName: \"kubernetes.io/projected/e28fb330-9e1a-4480-9a2d-b059d8e1576d-kube-api-access-xpnxr\") pod \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.505135 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-combined-ca-bundle\") pod \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\" (UID: \"e28fb330-9e1a-4480-9a2d-b059d8e1576d\") " Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.518181 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e28fb330-9e1a-4480-9a2d-b059d8e1576d" (UID: "e28fb330-9e1a-4480-9a2d-b059d8e1576d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.526673 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-scripts" (OuterVolumeSpecName: "scripts") pod "e28fb330-9e1a-4480-9a2d-b059d8e1576d" (UID: "e28fb330-9e1a-4480-9a2d-b059d8e1576d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.529271 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e28fb330-9e1a-4480-9a2d-b059d8e1576d" (UID: "e28fb330-9e1a-4480-9a2d-b059d8e1576d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.538519 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e28fb330-9e1a-4480-9a2d-b059d8e1576d-kube-api-access-xpnxr" (OuterVolumeSpecName: "kube-api-access-xpnxr") pod "e28fb330-9e1a-4480-9a2d-b059d8e1576d" (UID: "e28fb330-9e1a-4480-9a2d-b059d8e1576d"). InnerVolumeSpecName "kube-api-access-xpnxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.563549 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e28fb330-9e1a-4480-9a2d-b059d8e1576d" (UID: "e28fb330-9e1a-4480-9a2d-b059d8e1576d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.617918 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-config-data" (OuterVolumeSpecName: "config-data") pod "e28fb330-9e1a-4480-9a2d-b059d8e1576d" (UID: "e28fb330-9e1a-4480-9a2d-b059d8e1576d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.618893 4573 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.618940 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.618949 4573 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.618960 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.618970 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpnxr\" (UniqueName: \"kubernetes.io/projected/e28fb330-9e1a-4480-9a2d-b059d8e1576d-kube-api-access-xpnxr\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.618980 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28fb330-9e1a-4480-9a2d-b059d8e1576d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.650399 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m4vz4" event={"ID":"e28fb330-9e1a-4480-9a2d-b059d8e1576d","Type":"ContainerDied","Data":"7f58726d132b1790abcb1296dac03d02436caae1e6e9b42653d6e1c3ea7740b7"} Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.650446 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f58726d132b1790abcb1296dac03d02436caae1e6e9b42653d6e1c3ea7740b7" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.650516 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m4vz4" Dec 03 09:01:16 crc kubenswrapper[4573]: I1203 09:01:16.952405 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85bbcf664-5fm4n"] Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.668086 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8","Type":"ContainerStarted","Data":"90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02"} Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.669478 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85bbcf664-5fm4n" event={"ID":"b292e81d-5ded-4fff-bb9e-c764c395609b","Type":"ContainerStarted","Data":"c491735ed67f7e39c04b17a45c77e0c185c9be94d267404a84d6012181c8da24"} Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.686130 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/2.log" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.689671 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/1.log" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.692682 4573 generic.go:334] "Generic (PLEG): container finished" podID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerID="2e23fe287b388e79d38244fa35dc26ef66c81529d6a56def1aa3af463f92177c" exitCode=1 Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.692784 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerDied","Data":"2e23fe287b388e79d38244fa35dc26ef66c81529d6a56def1aa3af463f92177c"} Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.692852 4573 scope.go:117] "RemoveContainer" containerID="c356ec893e6d00cfbf39cf9f08708310b04998f0f0aa8654a1b0daa62c4db295" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.693866 4573 scope.go:117] "RemoveContainer" containerID="2e23fe287b388e79d38244fa35dc26ef66c81529d6a56def1aa3af463f92177c" Dec 03 09:01:17 crc kubenswrapper[4573]: E1203 09:01:17.694296 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 20s restarting failed container=neutron-httpd pod=neutron-756bc5c75b-vdw4l_openstack(574fcf75-461a-4cfa-b1e0-d9759ee7be40)\"" pod="openstack/neutron-756bc5c75b-vdw4l" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.705798 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fnppb" event={"ID":"3cd84894-b4fa-481f-8856-678218f7bad7","Type":"ContainerStarted","Data":"52e20e2bad1b07d9d79d83194aac0f7db5989ed58ba13e71f2fc4e566ec38b26"} Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.837971 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f56bd7d55-5qgxb"] Dec 03 09:01:17 crc kubenswrapper[4573]: E1203 09:01:17.838574 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e28fb330-9e1a-4480-9a2d-b059d8e1576d" containerName="keystone-bootstrap" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.838589 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e28fb330-9e1a-4480-9a2d-b059d8e1576d" containerName="keystone-bootstrap" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.838811 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="e28fb330-9e1a-4480-9a2d-b059d8e1576d" containerName="keystone-bootstrap" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.839680 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.852149 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.852447 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.852658 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.852815 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.855351 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.855482 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5wmq4" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.898764 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-fnppb" podStartSLOduration=18.4405417 podStartE2EDuration="1m18.898713897s" podCreationTimestamp="2025-12-03 08:59:59 +0000 UTC" firstStartedPulling="2025-12-03 09:00:04.376222244 +0000 UTC m=+1324.944601503" lastFinishedPulling="2025-12-03 09:01:04.834394441 +0000 UTC m=+1385.402773700" observedRunningTime="2025-12-03 09:01:17.802331533 +0000 UTC m=+1398.370710792" watchObservedRunningTime="2025-12-03 09:01:17.898713897 +0000 UTC m=+1398.467093156" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.955556 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f56bd7d55-5qgxb"] Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.963435 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-credential-keys\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.963487 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-scripts\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.963512 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-fernet-keys\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.963535 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwr2j\" (UniqueName: \"kubernetes.io/projected/06a92024-3b18-43c5-a4aa-22cec1d43ff3-kube-api-access-cwr2j\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.963584 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-config-data\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.963623 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-internal-tls-certs\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.963654 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-public-tls-certs\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:17 crc kubenswrapper[4573]: I1203 09:01:17.963676 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-combined-ca-bundle\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.068437 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-public-tls-certs\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.068495 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-combined-ca-bundle\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.068595 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-credential-keys\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.068614 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-scripts\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.068649 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-fernet-keys\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.068673 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwr2j\" (UniqueName: \"kubernetes.io/projected/06a92024-3b18-43c5-a4aa-22cec1d43ff3-kube-api-access-cwr2j\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.068740 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-config-data\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.068784 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-internal-tls-certs\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.079316 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-public-tls-certs\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.081525 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-combined-ca-bundle\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.083468 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-scripts\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.084817 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-fernet-keys\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.095793 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-credential-keys\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.096603 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-internal-tls-certs\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.110159 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06a92024-3b18-43c5-a4aa-22cec1d43ff3-config-data\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.130165 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwr2j\" (UniqueName: \"kubernetes.io/projected/06a92024-3b18-43c5-a4aa-22cec1d43ff3-kube-api-access-cwr2j\") pod \"keystone-f56bd7d55-5qgxb\" (UID: \"06a92024-3b18-43c5-a4aa-22cec1d43ff3\") " pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.202112 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.727318 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8f33c171-b5c5-444e-9cd0-63e2a2322f1c","Type":"ContainerStarted","Data":"5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1"} Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.729708 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-chs2b" event={"ID":"5c73103b-e4fd-4a54-ad2c-021db8042b13","Type":"ContainerStarted","Data":"d3a2b62460461d5aace5db31648f52a51f42f6610b73ce62ce94612ef769eab8"} Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.731511 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85bbcf664-5fm4n" event={"ID":"b292e81d-5ded-4fff-bb9e-c764c395609b","Type":"ContainerStarted","Data":"66c192638729a6b37449012954d7887aaacf6151e5723ad3fda1ed7fcf39e8cf"} Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.745198 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/2.log" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.772482 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=17.772432385 podStartE2EDuration="17.772432385s" podCreationTimestamp="2025-12-03 09:01:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:18.760943259 +0000 UTC m=+1399.329322518" watchObservedRunningTime="2025-12-03 09:01:18.772432385 +0000 UTC m=+1399.340811654" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.797286 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-chs2b" podStartSLOduration=18.457233502 podStartE2EDuration="1m19.797263559s" podCreationTimestamp="2025-12-03 08:59:59 +0000 UTC" firstStartedPulling="2025-12-03 09:00:03.494367674 +0000 UTC m=+1324.062746933" lastFinishedPulling="2025-12-03 09:01:04.834397731 +0000 UTC m=+1385.402776990" observedRunningTime="2025-12-03 09:01:18.793806904 +0000 UTC m=+1399.362186163" watchObservedRunningTime="2025-12-03 09:01:18.797263559 +0000 UTC m=+1399.365642818" Dec 03 09:01:18 crc kubenswrapper[4573]: I1203 09:01:18.946697 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f56bd7d55-5qgxb"] Dec 03 09:01:18 crc kubenswrapper[4573]: W1203 09:01:18.960544 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06a92024_3b18_43c5_a4aa_22cec1d43ff3.slice/crio-ea9e06e97e14044375fd6d16a361931b35d1852f38876ed53585de3879deb964 WatchSource:0}: Error finding container ea9e06e97e14044375fd6d16a361931b35d1852f38876ed53585de3879deb964: Status 404 returned error can't find the container with id ea9e06e97e14044375fd6d16a361931b35d1852f38876ed53585de3879deb964 Dec 03 09:01:19 crc kubenswrapper[4573]: I1203 09:01:19.780517 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85bbcf664-5fm4n" event={"ID":"b292e81d-5ded-4fff-bb9e-c764c395609b","Type":"ContainerStarted","Data":"fb648e2083aeaa075bc078ab8e580b74ea016b76f4dc1c1f1833c5c58b70639f"} Dec 03 09:01:19 crc kubenswrapper[4573]: I1203 09:01:19.782405 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:19 crc kubenswrapper[4573]: I1203 09:01:19.782449 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:19 crc kubenswrapper[4573]: I1203 09:01:19.790268 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f56bd7d55-5qgxb" event={"ID":"06a92024-3b18-43c5-a4aa-22cec1d43ff3","Type":"ContainerStarted","Data":"f59e1b4436a57fad74e6d7af2bd11027c90c44ea2da6223a31014a38e116c38e"} Dec 03 09:01:19 crc kubenswrapper[4573]: I1203 09:01:19.790312 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f56bd7d55-5qgxb" event={"ID":"06a92024-3b18-43c5-a4aa-22cec1d43ff3","Type":"ContainerStarted","Data":"ea9e06e97e14044375fd6d16a361931b35d1852f38876ed53585de3879deb964"} Dec 03 09:01:19 crc kubenswrapper[4573]: I1203 09:01:19.790330 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:19 crc kubenswrapper[4573]: I1203 09:01:19.823923 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-85bbcf664-5fm4n" podStartSLOduration=7.823903467 podStartE2EDuration="7.823903467s" podCreationTimestamp="2025-12-03 09:01:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:19.819338441 +0000 UTC m=+1400.387717700" watchObservedRunningTime="2025-12-03 09:01:19.823903467 +0000 UTC m=+1400.392282726" Dec 03 09:01:19 crc kubenswrapper[4573]: I1203 09:01:19.881519 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-f56bd7d55-5qgxb" podStartSLOduration=2.881494472 podStartE2EDuration="2.881494472s" podCreationTimestamp="2025-12-03 09:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:19.850796927 +0000 UTC m=+1400.419176186" watchObservedRunningTime="2025-12-03 09:01:19.881494472 +0000 UTC m=+1400.449873731" Dec 03 09:01:20 crc kubenswrapper[4573]: I1203 09:01:20.670453 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:20 crc kubenswrapper[4573]: I1203 09:01:20.670758 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:20 crc kubenswrapper[4573]: I1203 09:01:20.741742 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:20 crc kubenswrapper[4573]: I1203 09:01:20.749528 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:20 crc kubenswrapper[4573]: I1203 09:01:20.807568 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:20 crc kubenswrapper[4573]: I1203 09:01:20.807616 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:20 crc kubenswrapper[4573]: I1203 09:01:20.919476 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 09:01:21 crc kubenswrapper[4573]: I1203 09:01:21.072013 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 03 09:01:22 crc kubenswrapper[4573]: I1203 09:01:22.786708 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 09:01:22 crc kubenswrapper[4573]: I1203 09:01:22.787162 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 09:01:22 crc kubenswrapper[4573]: I1203 09:01:22.855011 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 09:01:22 crc kubenswrapper[4573]: I1203 09:01:22.858010 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 09:01:22 crc kubenswrapper[4573]: I1203 09:01:22.858102 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 09:01:22 crc kubenswrapper[4573]: I1203 09:01:22.858415 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 09:01:22 crc kubenswrapper[4573]: I1203 09:01:22.859558 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 09:01:23 crc kubenswrapper[4573]: I1203 09:01:23.867655 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 09:01:24 crc kubenswrapper[4573]: I1203 09:01:24.306258 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:01:24 crc kubenswrapper[4573]: I1203 09:01:24.306963 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:01:24 crc kubenswrapper[4573]: I1203 09:01:24.306997 4573 scope.go:117] "RemoveContainer" containerID="2e23fe287b388e79d38244fa35dc26ef66c81529d6a56def1aa3af463f92177c" Dec 03 09:01:24 crc kubenswrapper[4573]: E1203 09:01:24.307379 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 20s restarting failed container=neutron-httpd pod=neutron-756bc5c75b-vdw4l_openstack(574fcf75-461a-4cfa-b1e0-d9759ee7be40)\"" pod="openstack/neutron-756bc5c75b-vdw4l" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" Dec 03 09:01:24 crc kubenswrapper[4573]: I1203 09:01:24.312253 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-756bc5c75b-vdw4l" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-api" probeResult="failure" output="Get \"http://10.217.0.151:9696/\": dial tcp 10.217.0.151:9696: connect: connection refused" Dec 03 09:01:24 crc kubenswrapper[4573]: I1203 09:01:24.914480 4573 generic.go:334] "Generic (PLEG): container finished" podID="3cd84894-b4fa-481f-8856-678218f7bad7" containerID="52e20e2bad1b07d9d79d83194aac0f7db5989ed58ba13e71f2fc4e566ec38b26" exitCode=0 Dec 03 09:01:24 crc kubenswrapper[4573]: I1203 09:01:24.915818 4573 scope.go:117] "RemoveContainer" containerID="2e23fe287b388e79d38244fa35dc26ef66c81529d6a56def1aa3af463f92177c" Dec 03 09:01:24 crc kubenswrapper[4573]: E1203 09:01:24.916120 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 20s restarting failed container=neutron-httpd pod=neutron-756bc5c75b-vdw4l_openstack(574fcf75-461a-4cfa-b1e0-d9759ee7be40)\"" pod="openstack/neutron-756bc5c75b-vdw4l" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" Dec 03 09:01:24 crc kubenswrapper[4573]: I1203 09:01:24.917431 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fnppb" event={"ID":"3cd84894-b4fa-481f-8856-678218f7bad7","Type":"ContainerDied","Data":"52e20e2bad1b07d9d79d83194aac0f7db5989ed58ba13e71f2fc4e566ec38b26"} Dec 03 09:01:24 crc kubenswrapper[4573]: I1203 09:01:24.917993 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 09:01:26 crc kubenswrapper[4573]: I1203 09:01:26.946849 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:01:26 crc kubenswrapper[4573]: I1203 09:01:26.947564 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:01:27 crc kubenswrapper[4573]: I1203 09:01:27.237464 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:27 crc kubenswrapper[4573]: I1203 09:01:27.237583 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 09:01:27 crc kubenswrapper[4573]: I1203 09:01:27.248747 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 09:01:27 crc kubenswrapper[4573]: I1203 09:01:27.248887 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 09:01:27 crc kubenswrapper[4573]: I1203 09:01:27.252218 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 09:01:27 crc kubenswrapper[4573]: I1203 09:01:27.817902 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 09:01:29 crc kubenswrapper[4573]: I1203 09:01:29.116817 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cfccd4555-2shp6" Dec 03 09:01:29 crc kubenswrapper[4573]: I1203 09:01:29.493883 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-756bc5c75b-vdw4l"] Dec 03 09:01:29 crc kubenswrapper[4573]: I1203 09:01:29.497114 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-756bc5c75b-vdw4l" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-api" containerID="cri-o://cb8dd971f575f2e4c8253374b56b43818da2c0e0bb78fb0073202f4cd1f5eb47" gracePeriod=30 Dec 03 09:01:30 crc kubenswrapper[4573]: I1203 09:01:30.919461 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 09:01:30 crc kubenswrapper[4573]: I1203 09:01:30.919938 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:01:30 crc kubenswrapper[4573]: I1203 09:01:30.921191 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"875e1f86deecceeb9c350e1909663db37b4f6f0a82cff5c1165cad17d9cdc1df"} pod="openstack/horizon-7b8b5b6556-4zxxs" containerMessage="Container horizon failed startup probe, will be restarted" Dec 03 09:01:30 crc kubenswrapper[4573]: I1203 09:01:30.921246 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" containerID="cri-o://875e1f86deecceeb9c350e1909663db37b4f6f0a82cff5c1165cad17d9cdc1df" gracePeriod=30 Dec 03 09:01:31 crc kubenswrapper[4573]: I1203 09:01:31.071407 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 03 09:01:31 crc kubenswrapper[4573]: I1203 09:01:31.071750 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:01:31 crc kubenswrapper[4573]: I1203 09:01:31.072639 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"afd0f349dd0be138a212ad8426fb19507ce3fcb56dbe6dbd953d89456cd91458"} pod="openstack/horizon-66f99c579b-gh7sh" containerMessage="Container horizon failed startup probe, will be restarted" Dec 03 09:01:31 crc kubenswrapper[4573]: I1203 09:01:31.072771 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" containerID="cri-o://afd0f349dd0be138a212ad8426fb19507ce3fcb56dbe6dbd953d89456cd91458" gracePeriod=30 Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.113090 4573 generic.go:334] "Generic (PLEG): container finished" podID="5c73103b-e4fd-4a54-ad2c-021db8042b13" containerID="d3a2b62460461d5aace5db31648f52a51f42f6610b73ce62ce94612ef769eab8" exitCode=0 Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.113196 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-chs2b" event={"ID":"5c73103b-e4fd-4a54-ad2c-021db8042b13","Type":"ContainerDied","Data":"d3a2b62460461d5aace5db31648f52a51f42f6610b73ce62ce94612ef769eab8"} Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.137507 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/2.log" Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.139211 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerDied","Data":"cb8dd971f575f2e4c8253374b56b43818da2c0e0bb78fb0073202f4cd1f5eb47"} Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.142399 4573 generic.go:334] "Generic (PLEG): container finished" podID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerID="cb8dd971f575f2e4c8253374b56b43818da2c0e0bb78fb0073202f4cd1f5eb47" exitCode=0 Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.376311 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fnppb" Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.485541 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxl4r\" (UniqueName: \"kubernetes.io/projected/3cd84894-b4fa-481f-8856-678218f7bad7-kube-api-access-xxl4r\") pod \"3cd84894-b4fa-481f-8856-678218f7bad7\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.485664 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-db-sync-config-data\") pod \"3cd84894-b4fa-481f-8856-678218f7bad7\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.485730 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-combined-ca-bundle\") pod \"3cd84894-b4fa-481f-8856-678218f7bad7\" (UID: \"3cd84894-b4fa-481f-8856-678218f7bad7\") " Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.521390 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd84894-b4fa-481f-8856-678218f7bad7-kube-api-access-xxl4r" (OuterVolumeSpecName: "kube-api-access-xxl4r") pod "3cd84894-b4fa-481f-8856-678218f7bad7" (UID: "3cd84894-b4fa-481f-8856-678218f7bad7"). InnerVolumeSpecName "kube-api-access-xxl4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.521550 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3cd84894-b4fa-481f-8856-678218f7bad7" (UID: "3cd84894-b4fa-481f-8856-678218f7bad7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.537151 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cd84894-b4fa-481f-8856-678218f7bad7" (UID: "3cd84894-b4fa-481f-8856-678218f7bad7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.589416 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.589455 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxl4r\" (UniqueName: \"kubernetes.io/projected/3cd84894-b4fa-481f-8856-678218f7bad7-kube-api-access-xxl4r\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:34 crc kubenswrapper[4573]: I1203 09:01:34.589468 4573 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3cd84894-b4fa-481f-8856-678218f7bad7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.167241 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-fnppb" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.167917 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-fnppb" event={"ID":"3cd84894-b4fa-481f-8856-678218f7bad7","Type":"ContainerDied","Data":"5fe8c07896c62a9c99c258f7e2c8e1a57064244015771c1cac217a86a080bb78"} Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.167996 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fe8c07896c62a9c99c258f7e2c8e1a57064244015771c1cac217a86a080bb78" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.787456 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-cd9d86d7-kn6nh"] Dec 03 09:01:35 crc kubenswrapper[4573]: E1203 09:01:35.787967 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd84894-b4fa-481f-8856-678218f7bad7" containerName="barbican-db-sync" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.787990 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd84894-b4fa-481f-8856-678218f7bad7" containerName="barbican-db-sync" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.788245 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd84894-b4fa-481f-8856-678218f7bad7" containerName="barbican-db-sync" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.801917 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.808865 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jkqdd" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.809309 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.809469 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.841406 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-cd9d86d7-kn6nh"] Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.870146 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5c965d87c4-dmtfc"] Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.876497 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.885468 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.902099 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5c965d87c4-dmtfc"] Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.925585 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fb2z\" (UniqueName: \"kubernetes.io/projected/865e87dd-1725-4932-9566-dbfbf7b85a60-kube-api-access-2fb2z\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.925677 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-combined-ca-bundle\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.925730 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/865e87dd-1725-4932-9566-dbfbf7b85a60-logs\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.925839 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-config-data-custom\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:35 crc kubenswrapper[4573]: I1203 09:01:35.925893 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-config-data\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.034973 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-config-data\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035082 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-config-data-custom\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035132 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fb2z\" (UniqueName: \"kubernetes.io/projected/865e87dd-1725-4932-9566-dbfbf7b85a60-kube-api-access-2fb2z\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035172 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-config-data\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035202 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gwr2\" (UniqueName: \"kubernetes.io/projected/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-kube-api-access-6gwr2\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035236 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-combined-ca-bundle\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035260 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/865e87dd-1725-4932-9566-dbfbf7b85a60-logs\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035282 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-combined-ca-bundle\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035298 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-logs\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.035350 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-config-data-custom\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.036835 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/865e87dd-1725-4932-9566-dbfbf7b85a60-logs\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.057713 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-combined-ca-bundle\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.058330 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-config-data-custom\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.074790 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/865e87dd-1725-4932-9566-dbfbf7b85a60-config-data\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.091885 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fb2z\" (UniqueName: \"kubernetes.io/projected/865e87dd-1725-4932-9566-dbfbf7b85a60-kube-api-access-2fb2z\") pod \"barbican-worker-cd9d86d7-kn6nh\" (UID: \"865e87dd-1725-4932-9566-dbfbf7b85a60\") " pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.112911 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pblbw"] Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.115241 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.137208 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-combined-ca-bundle\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.137274 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-logs\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.146581 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-logs\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.146860 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-combined-ca-bundle\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.150325 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-config-data-custom\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.150516 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-config-data\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.150551 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gwr2\" (UniqueName: \"kubernetes.io/projected/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-kube-api-access-6gwr2\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.163227 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pblbw"] Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.171379 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-cd9d86d7-kn6nh" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.183577 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-config-data\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.190425 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-config-data-custom\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.225258 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gwr2\" (UniqueName: \"kubernetes.io/projected/e954edc8-66c8-4f8f-a255-b8c5d810aa1a-kube-api-access-6gwr2\") pod \"barbican-keystone-listener-5c965d87c4-dmtfc\" (UID: \"e954edc8-66c8-4f8f-a255-b8c5d810aa1a\") " pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.258902 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.272603 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6fbfd76774-v2hm7"] Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.300057 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-config\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.301390 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.306516 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.306576 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.306867 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.314274 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjh6v\" (UniqueName: \"kubernetes.io/projected/27e16259-d436-411f-97af-e8ff9a8be41c-kube-api-access-hjh6v\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.329663 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.334952 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.420394 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.421003 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a37c63f1-b984-4c85-9866-e54541dc09b8-logs\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.422749 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428022 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428143 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428300 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz48z\" (UniqueName: \"kubernetes.io/projected/a37c63f1-b984-4c85-9866-e54541dc09b8-kube-api-access-kz48z\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428383 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data-custom\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428450 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428477 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-combined-ca-bundle\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428544 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428571 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjh6v\" (UniqueName: \"kubernetes.io/projected/27e16259-d436-411f-97af-e8ff9a8be41c-kube-api-access-hjh6v\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.428663 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-config\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.429940 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.430762 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.431376 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.448107 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-config\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.456116 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjh6v\" (UniqueName: \"kubernetes.io/projected/27e16259-d436-411f-97af-e8ff9a8be41c-kube-api-access-hjh6v\") pod \"dnsmasq-dns-848cf88cfc-pblbw\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.454392 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6fbfd76774-v2hm7"] Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.531610 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz48z\" (UniqueName: \"kubernetes.io/projected/a37c63f1-b984-4c85-9866-e54541dc09b8-kube-api-access-kz48z\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.531695 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data-custom\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.531728 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.531774 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-combined-ca-bundle\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.531976 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a37c63f1-b984-4c85-9866-e54541dc09b8-logs\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.532768 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a37c63f1-b984-4c85-9866-e54541dc09b8-logs\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.540899 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data-custom\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.545928 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.558306 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz48z\" (UniqueName: \"kubernetes.io/projected/a37c63f1-b984-4c85-9866-e54541dc09b8-kube-api-access-kz48z\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.563815 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-combined-ca-bundle\") pod \"barbican-api-6fbfd76774-v2hm7\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.676355 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:36 crc kubenswrapper[4573]: I1203 09:01:36.726001 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.012999 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-chs2b" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.204529 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-db-sync-config-data\") pod \"5c73103b-e4fd-4a54-ad2c-021db8042b13\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.204705 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwfgd\" (UniqueName: \"kubernetes.io/projected/5c73103b-e4fd-4a54-ad2c-021db8042b13-kube-api-access-vwfgd\") pod \"5c73103b-e4fd-4a54-ad2c-021db8042b13\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.205006 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-combined-ca-bundle\") pod \"5c73103b-e4fd-4a54-ad2c-021db8042b13\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.205042 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-scripts\") pod \"5c73103b-e4fd-4a54-ad2c-021db8042b13\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.205104 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c73103b-e4fd-4a54-ad2c-021db8042b13-etc-machine-id\") pod \"5c73103b-e4fd-4a54-ad2c-021db8042b13\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.205143 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-config-data\") pod \"5c73103b-e4fd-4a54-ad2c-021db8042b13\" (UID: \"5c73103b-e4fd-4a54-ad2c-021db8042b13\") " Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.213184 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c73103b-e4fd-4a54-ad2c-021db8042b13-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5c73103b-e4fd-4a54-ad2c-021db8042b13" (UID: "5c73103b-e4fd-4a54-ad2c-021db8042b13"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.226666 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5c73103b-e4fd-4a54-ad2c-021db8042b13" (UID: "5c73103b-e4fd-4a54-ad2c-021db8042b13"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.252564 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c73103b-e4fd-4a54-ad2c-021db8042b13-kube-api-access-vwfgd" (OuterVolumeSpecName: "kube-api-access-vwfgd") pod "5c73103b-e4fd-4a54-ad2c-021db8042b13" (UID: "5c73103b-e4fd-4a54-ad2c-021db8042b13"). InnerVolumeSpecName "kube-api-access-vwfgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.293373 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-scripts" (OuterVolumeSpecName: "scripts") pod "5c73103b-e4fd-4a54-ad2c-021db8042b13" (UID: "5c73103b-e4fd-4a54-ad2c-021db8042b13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.297252 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c73103b-e4fd-4a54-ad2c-021db8042b13" (UID: "5c73103b-e4fd-4a54-ad2c-021db8042b13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.320706 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.321031 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.321135 4573 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c73103b-e4fd-4a54-ad2c-021db8042b13-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.321223 4573 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.321326 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwfgd\" (UniqueName: \"kubernetes.io/projected/5c73103b-e4fd-4a54-ad2c-021db8042b13-kube-api-access-vwfgd\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.374350 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-chs2b" event={"ID":"5c73103b-e4fd-4a54-ad2c-021db8042b13","Type":"ContainerDied","Data":"01feccc9862109d0a879828bf169aa6fb844d8f0240db330650b322ea3407d7f"} Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.374408 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01feccc9862109d0a879828bf169aa6fb844d8f0240db330650b322ea3407d7f" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.374507 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-chs2b" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.393629 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-config-data" (OuterVolumeSpecName: "config-data") pod "5c73103b-e4fd-4a54-ad2c-021db8042b13" (UID: "5c73103b-e4fd-4a54-ad2c-021db8042b13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:37 crc kubenswrapper[4573]: I1203 09:01:37.428953 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c73103b-e4fd-4a54-ad2c-021db8042b13-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.347840 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:01:38 crc kubenswrapper[4573]: E1203 09:01:38.348847 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c73103b-e4fd-4a54-ad2c-021db8042b13" containerName="cinder-db-sync" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.348869 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c73103b-e4fd-4a54-ad2c-021db8042b13" containerName="cinder-db-sync" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.349124 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c73103b-e4fd-4a54-ad2c-021db8042b13" containerName="cinder-db-sync" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.352222 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.369672 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.374032 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-qmt4b" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.374438 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.374720 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.414439 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.457813 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.457875 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-scripts\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.457946 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c1c9237-25b7-41d3-bfc6-ac8dea228059-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.457994 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8fpc\" (UniqueName: \"kubernetes.io/projected/8c1c9237-25b7-41d3-bfc6-ac8dea228059-kube-api-access-l8fpc\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.458038 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.458144 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.530342 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pblbw"] Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.560781 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8fpc\" (UniqueName: \"kubernetes.io/projected/8c1c9237-25b7-41d3-bfc6-ac8dea228059-kube-api-access-l8fpc\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.560866 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.560947 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.561011 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.561041 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-scripts\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.561108 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c1c9237-25b7-41d3-bfc6-ac8dea228059-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.561294 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c1c9237-25b7-41d3-bfc6-ac8dea228059-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.588494 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.596642 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.601167 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-scripts\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.608232 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8fpc\" (UniqueName: \"kubernetes.io/projected/8c1c9237-25b7-41d3-bfc6-ac8dea228059-kube-api-access-l8fpc\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.616299 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-79ggx"] Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.618485 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.687739 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.688617 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.732143 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-79ggx"] Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.773421 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.773568 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-config\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.773641 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-svc\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.773718 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.773768 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hnmg\" (UniqueName: \"kubernetes.io/projected/b9a834f3-5601-48b5-982a-0ab476605db5-kube-api-access-4hnmg\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.773828 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.850418 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.856391 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.861992 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.868472 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.876946 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-config\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.877022 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-svc\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.877082 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.877121 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hnmg\" (UniqueName: \"kubernetes.io/projected/b9a834f3-5601-48b5-982a-0ab476605db5-kube-api-access-4hnmg\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.877171 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.877212 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.878493 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.882302 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.885201 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.886136 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-config\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.887979 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-svc\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.919609 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hnmg\" (UniqueName: \"kubernetes.io/projected/b9a834f3-5601-48b5-982a-0ab476605db5-kube-api-access-4hnmg\") pod \"dnsmasq-dns-6578955fd5-79ggx\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.979722 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.979887 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04646290-c4ca-444e-b752-12a054221b16-logs\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.979971 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-scripts\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.980084 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04646290-c4ca-444e-b752-12a054221b16-etc-machine-id\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.980177 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6sll\" (UniqueName: \"kubernetes.io/projected/04646290-c4ca-444e-b752-12a054221b16-kube-api-access-g6sll\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.980249 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data-custom\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:38 crc kubenswrapper[4573]: I1203 09:01:38.980332 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: E1203 09:01:39.072958 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Dec 03 09:01:39 crc kubenswrapper[4573]: E1203 09:01:39.073252 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rb5sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 09:01:39 crc kubenswrapper[4573]: E1203 09:01:39.075334 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"ceilometer-notification-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.083137 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.083231 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.083275 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04646290-c4ca-444e-b752-12a054221b16-logs\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.083381 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-scripts\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.083505 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04646290-c4ca-444e-b752-12a054221b16-etc-machine-id\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.083647 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6sll\" (UniqueName: \"kubernetes.io/projected/04646290-c4ca-444e-b752-12a054221b16-kube-api-access-g6sll\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.083735 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data-custom\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.085709 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04646290-c4ca-444e-b752-12a054221b16-logs\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.086847 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04646290-c4ca-444e-b752-12a054221b16-etc-machine-id\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.089376 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.089794 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data-custom\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.096176 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.102806 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.104729 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-scripts\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.112798 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6sll\" (UniqueName: \"kubernetes.io/projected/04646290-c4ca-444e-b752-12a054221b16-kube-api-access-g6sll\") pod \"cinder-api-0\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.130468 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/2.log" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.132313 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.200665 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.286978 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-config\") pod \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.287099 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqlxt\" (UniqueName: \"kubernetes.io/projected/574fcf75-461a-4cfa-b1e0-d9759ee7be40-kube-api-access-nqlxt\") pod \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.287248 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-ovndb-tls-certs\") pod \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.287346 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-httpd-config\") pod \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.287444 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-combined-ca-bundle\") pod \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\" (UID: \"574fcf75-461a-4cfa-b1e0-d9759ee7be40\") " Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.293408 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574fcf75-461a-4cfa-b1e0-d9759ee7be40-kube-api-access-nqlxt" (OuterVolumeSpecName: "kube-api-access-nqlxt") pod "574fcf75-461a-4cfa-b1e0-d9759ee7be40" (UID: "574fcf75-461a-4cfa-b1e0-d9759ee7be40"). InnerVolumeSpecName "kube-api-access-nqlxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.304005 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "574fcf75-461a-4cfa-b1e0-d9759ee7be40" (UID: "574fcf75-461a-4cfa-b1e0-d9759ee7be40"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.393494 4573 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.393785 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqlxt\" (UniqueName: \"kubernetes.io/projected/574fcf75-461a-4cfa-b1e0-d9759ee7be40-kube-api-access-nqlxt\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.483086 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-config" (OuterVolumeSpecName: "config") pod "574fcf75-461a-4cfa-b1e0-d9759ee7be40" (UID: "574fcf75-461a-4cfa-b1e0-d9759ee7be40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.484276 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-756bc5c75b-vdw4l_574fcf75-461a-4cfa-b1e0-d9759ee7be40/neutron-httpd/2.log" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.487122 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" containerName="sg-core" containerID="cri-o://90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02" gracePeriod=30 Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.487215 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756bc5c75b-vdw4l" event={"ID":"574fcf75-461a-4cfa-b1e0-d9759ee7be40","Type":"ContainerDied","Data":"9f2c13ff514b3a75bc91efb9fbc98c8ab8b6501081e3b9794298cb1e4fdc4af3"} Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.487289 4573 scope.go:117] "RemoveContainer" containerID="2e23fe287b388e79d38244fa35dc26ef66c81529d6a56def1aa3af463f92177c" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.487161 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756bc5c75b-vdw4l" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.496153 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.541994 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "574fcf75-461a-4cfa-b1e0-d9759ee7be40" (UID: "574fcf75-461a-4cfa-b1e0-d9759ee7be40"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.555538 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "574fcf75-461a-4cfa-b1e0-d9759ee7be40" (UID: "574fcf75-461a-4cfa-b1e0-d9759ee7be40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.629575 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.629608 4573 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/574fcf75-461a-4cfa-b1e0-d9759ee7be40-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:39 crc kubenswrapper[4573]: I1203 09:01:39.914792 4573 scope.go:117] "RemoveContainer" containerID="cb8dd971f575f2e4c8253374b56b43818da2c0e0bb78fb0073202f4cd1f5eb47" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.111128 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-756bc5c75b-vdw4l"] Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.150368 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-756bc5c75b-vdw4l"] Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.439737 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pblbw"] Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.599297 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" event={"ID":"27e16259-d436-411f-97af-e8ff9a8be41c","Type":"ContainerStarted","Data":"e4ee22dc627efc48c9f28e12cb210a41a00d868af65194d4376c49d3f8b465df"} Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.601155 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.619516 4573 generic.go:334] "Generic (PLEG): container finished" podID="ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" containerID="90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02" exitCode=2 Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.619591 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8","Type":"ContainerDied","Data":"90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02"} Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.639806 4573 scope.go:117] "RemoveContainer" containerID="90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.639762 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8","Type":"ContainerDied","Data":"ae957ce0eb270cee8fe71608b36de9ac14a47258e80a4c59094aead06b818a6e"} Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.677158 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-cd9d86d7-kn6nh"] Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.691839 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb5sz\" (UniqueName: \"kubernetes.io/projected/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-kube-api-access-rb5sz\") pod \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.691896 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-run-httpd\") pod \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.692006 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-config-data\") pod \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.692165 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-combined-ca-bundle\") pod \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.692282 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-scripts\") pod \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.692345 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-sg-core-conf-yaml\") pod \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.692392 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-log-httpd\") pod \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\" (UID: \"ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8\") " Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.693807 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" (UID: "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.701895 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-kube-api-access-rb5sz" (OuterVolumeSpecName: "kube-api-access-rb5sz") pod "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" (UID: "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8"). InnerVolumeSpecName "kube-api-access-rb5sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.705165 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" (UID: "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.736745 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" (UID: "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.746746 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-scripts" (OuterVolumeSpecName: "scripts") pod "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" (UID: "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.766315 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-config-data" (OuterVolumeSpecName: "config-data") pod "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" (UID: "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.768215 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6fbfd76774-v2hm7"] Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.795888 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.795935 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb5sz\" (UniqueName: \"kubernetes.io/projected/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-kube-api-access-rb5sz\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.795946 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.795957 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.795971 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.795980 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.962819 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.973909 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" (UID: "ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.976570 4573 scope.go:117] "RemoveContainer" containerID="90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02" Dec 03 09:01:40 crc kubenswrapper[4573]: E1203 09:01:40.982335 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02\": container with ID starting with 90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02 not found: ID does not exist" containerID="90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02" Dec 03 09:01:40 crc kubenswrapper[4573]: I1203 09:01:40.982412 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02"} err="failed to get container status \"90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02\": rpc error: code = NotFound desc = could not find container \"90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02\": container with ID starting with 90bd2ffe62e4eb487a351fc9c8cb860251002fef81c73dd3c2e8a4912991ca02 not found: ID does not exist" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.007468 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.258629 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.327978 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5c965d87c4-dmtfc"] Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.422542 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-79ggx"] Dec 03 09:01:41 crc kubenswrapper[4573]: W1203 09:01:41.424060 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9a834f3_5601_48b5_982a_0ab476605db5.slice/crio-f3800725ea7fe587487ff6f8a06d2381b616d0a05f1f555811bbab9da2653d8b WatchSource:0}: Error finding container f3800725ea7fe587487ff6f8a06d2381b616d0a05f1f555811bbab9da2653d8b: Status 404 returned error can't find the container with id f3800725ea7fe587487ff6f8a06d2381b616d0a05f1f555811bbab9da2653d8b Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.710398 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fbfd76774-v2hm7" event={"ID":"a37c63f1-b984-4c85-9866-e54541dc09b8","Type":"ContainerStarted","Data":"3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86"} Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.710909 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fbfd76774-v2hm7" event={"ID":"a37c63f1-b984-4c85-9866-e54541dc09b8","Type":"ContainerStarted","Data":"6326635287a540d0d0740b2fdf3c8534aee273f1953626ae766cf7eb1158c091"} Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.715311 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" event={"ID":"b9a834f3-5601-48b5-982a-0ab476605db5","Type":"ContainerStarted","Data":"f3800725ea7fe587487ff6f8a06d2381b616d0a05f1f555811bbab9da2653d8b"} Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.717228 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8c1c9237-25b7-41d3-bfc6-ac8dea228059","Type":"ContainerStarted","Data":"7418dab3f76d21a40e83486cfb1b258ccb71ab1e5b39d13b52eeeeb540e79d3c"} Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.721871 4573 generic.go:334] "Generic (PLEG): container finished" podID="27e16259-d436-411f-97af-e8ff9a8be41c" containerID="c5956299c8ea56cfdb151e5d1f8fedf175cd50471b40ea73f65dbb1b13e354da" exitCode=0 Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.721938 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" event={"ID":"27e16259-d436-411f-97af-e8ff9a8be41c","Type":"ContainerDied","Data":"c5956299c8ea56cfdb151e5d1f8fedf175cd50471b40ea73f65dbb1b13e354da"} Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.741175 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"04646290-c4ca-444e-b752-12a054221b16","Type":"ContainerStarted","Data":"77e9185a104bdca0d50f65af81ba892a2ec93290cee0019eaddb3e0933bff284"} Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.767332 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" event={"ID":"e954edc8-66c8-4f8f-a255-b8c5d810aa1a","Type":"ContainerStarted","Data":"12a63d24e1c5b11b7ff9418e0f2898fc67b96220ef8926829368b88ca90ee09d"} Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.774692 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-cd9d86d7-kn6nh" event={"ID":"865e87dd-1725-4932-9566-dbfbf7b85a60","Type":"ContainerStarted","Data":"228bae879dafa56e97340e875f5d01e2c524e3c27e1e66447d85c1f14f00c757"} Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.779110 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.894320 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.910528 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.925275 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:01:41 crc kubenswrapper[4573]: E1203 09:01:41.925871 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.925885 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: E1203 09:01:41.925901 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-api" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.925907 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-api" Dec 03 09:01:41 crc kubenswrapper[4573]: E1203 09:01:41.925930 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.925936 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: E1203 09:01:41.925946 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.925952 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: E1203 09:01:41.925972 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" containerName="sg-core" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.925978 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" containerName="sg-core" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.926152 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.926170 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" containerName="sg-core" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.926184 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.926195 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-api" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.926541 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" containerName="neutron-httpd" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.937808 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.950291 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.964616 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:01:41 crc kubenswrapper[4573]: I1203 09:01:41.997028 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.072812 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574fcf75-461a-4cfa-b1e0-d9759ee7be40" path="/var/lib/kubelet/pods/574fcf75-461a-4cfa-b1e0-d9759ee7be40/volumes" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.078790 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5fnd\" (UniqueName: \"kubernetes.io/projected/90faa05a-6a95-4261-957f-44b45ea20a02-kube-api-access-l5fnd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.078873 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-scripts\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.078948 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.079005 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-run-httpd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.079032 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-config-data\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.079069 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.079110 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-log-httpd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.080020 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8" path="/var/lib/kubelet/pods/ca8e17a9-f269-409e-ae1b-fc0aaaa22cc8/volumes" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.181061 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-run-httpd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.181134 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-config-data\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.181199 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.181255 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-log-httpd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.181910 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-run-httpd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.182358 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5fnd\" (UniqueName: \"kubernetes.io/projected/90faa05a-6a95-4261-957f-44b45ea20a02-kube-api-access-l5fnd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.182587 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-scripts\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.182761 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.187223 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-log-httpd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.188701 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-config-data\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.189743 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.197829 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-scripts\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.233159 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.246082 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5fnd\" (UniqueName: \"kubernetes.io/projected/90faa05a-6a95-4261-957f-44b45ea20a02-kube-api-access-l5fnd\") pod \"ceilometer-0\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.340126 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.724687 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.775128 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-swift-storage-0\") pod \"27e16259-d436-411f-97af-e8ff9a8be41c\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.818625 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "27e16259-d436-411f-97af-e8ff9a8be41c" (UID: "27e16259-d436-411f-97af-e8ff9a8be41c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.836917 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" event={"ID":"27e16259-d436-411f-97af-e8ff9a8be41c","Type":"ContainerDied","Data":"e4ee22dc627efc48c9f28e12cb210a41a00d868af65194d4376c49d3f8b465df"} Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.837009 4573 scope.go:117] "RemoveContainer" containerID="c5956299c8ea56cfdb151e5d1f8fedf175cd50471b40ea73f65dbb1b13e354da" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.838035 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-pblbw" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.880618 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-config\") pod \"27e16259-d436-411f-97af-e8ff9a8be41c\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.880763 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-sb\") pod \"27e16259-d436-411f-97af-e8ff9a8be41c\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.880871 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-svc\") pod \"27e16259-d436-411f-97af-e8ff9a8be41c\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.880968 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-nb\") pod \"27e16259-d436-411f-97af-e8ff9a8be41c\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.881078 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjh6v\" (UniqueName: \"kubernetes.io/projected/27e16259-d436-411f-97af-e8ff9a8be41c-kube-api-access-hjh6v\") pod \"27e16259-d436-411f-97af-e8ff9a8be41c\" (UID: \"27e16259-d436-411f-97af-e8ff9a8be41c\") " Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.881718 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.915108 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e16259-d436-411f-97af-e8ff9a8be41c-kube-api-access-hjh6v" (OuterVolumeSpecName: "kube-api-access-hjh6v") pod "27e16259-d436-411f-97af-e8ff9a8be41c" (UID: "27e16259-d436-411f-97af-e8ff9a8be41c"). InnerVolumeSpecName "kube-api-access-hjh6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:42 crc kubenswrapper[4573]: I1203 09:01:42.968958 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "27e16259-d436-411f-97af-e8ff9a8be41c" (UID: "27e16259-d436-411f-97af-e8ff9a8be41c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:42.988977 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:42.989058 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjh6v\" (UniqueName: \"kubernetes.io/projected/27e16259-d436-411f-97af-e8ff9a8be41c-kube-api-access-hjh6v\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:42.998320 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-config" (OuterVolumeSpecName: "config") pod "27e16259-d436-411f-97af-e8ff9a8be41c" (UID: "27e16259-d436-411f-97af-e8ff9a8be41c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.004735 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "27e16259-d436-411f-97af-e8ff9a8be41c" (UID: "27e16259-d436-411f-97af-e8ff9a8be41c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.009499 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fbfd76774-v2hm7" event={"ID":"a37c63f1-b984-4c85-9866-e54541dc09b8","Type":"ContainerStarted","Data":"175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316"} Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.012592 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.012632 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.037392 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "27e16259-d436-411f-97af-e8ff9a8be41c" (UID: "27e16259-d436-411f-97af-e8ff9a8be41c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.060373 4573 generic.go:334] "Generic (PLEG): container finished" podID="b9a834f3-5601-48b5-982a-0ab476605db5" containerID="82d3e11445e4fcc3789cf87aedc48577465aa73fc4f5f6e9b073d2282c3bab87" exitCode=0 Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.060463 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" event={"ID":"b9a834f3-5601-48b5-982a-0ab476605db5","Type":"ContainerDied","Data":"82d3e11445e4fcc3789cf87aedc48577465aa73fc4f5f6e9b073d2282c3bab87"} Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.079015 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6fbfd76774-v2hm7" podStartSLOduration=7.078988862 podStartE2EDuration="7.078988862s" podCreationTimestamp="2025-12-03 09:01:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:43.051910196 +0000 UTC m=+1423.620289465" watchObservedRunningTime="2025-12-03 09:01:43.078988862 +0000 UTC m=+1423.647368121" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.091689 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.091739 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.091754 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/27e16259-d436-411f-97af-e8ff9a8be41c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.326165 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pblbw"] Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.440277 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pblbw"] Dec 03 09:01:43 crc kubenswrapper[4573]: I1203 09:01:43.808306 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.054472 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27e16259-d436-411f-97af-e8ff9a8be41c" path="/var/lib/kubelet/pods/27e16259-d436-411f-97af-e8ff9a8be41c/volumes" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.100095 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" event={"ID":"b9a834f3-5601-48b5-982a-0ab476605db5","Type":"ContainerStarted","Data":"bb16388dcadba35686654bf424cad2e59956e899e5746a662a72cd0a6d693540"} Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.100166 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.108378 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"04646290-c4ca-444e-b752-12a054221b16","Type":"ContainerStarted","Data":"d86adc4513427e0feed45141efce0fd0fb33dafcb7ec082b101bd0c0e3acd93d"} Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.134225 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" podStartSLOduration=6.134187716 podStartE2EDuration="6.134187716s" podCreationTimestamp="2025-12-03 09:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:44.126790022 +0000 UTC m=+1424.695169361" watchObservedRunningTime="2025-12-03 09:01:44.134187716 +0000 UTC m=+1424.702566975" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.162160 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.903578 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-847b4dbb6-dsqmk"] Dec 03 09:01:44 crc kubenswrapper[4573]: E1203 09:01:44.904632 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e16259-d436-411f-97af-e8ff9a8be41c" containerName="init" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.904653 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e16259-d436-411f-97af-e8ff9a8be41c" containerName="init" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.904862 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e16259-d436-411f-97af-e8ff9a8be41c" containerName="init" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.913871 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.923636 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.923883 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 09:01:44 crc kubenswrapper[4573]: I1203 09:01:44.942293 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-847b4dbb6-dsqmk"] Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.167156 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8c1c9237-25b7-41d3-bfc6-ac8dea228059","Type":"ContainerStarted","Data":"35675082d40a9c2fd01697ae0adc20fc68defb5adb12ce6f6ea13227526d79c5"} Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.187186 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-config-data\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.187277 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-config-data-custom\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.187347 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6907cde0-9a5e-432e-bc68-55b357a514d9-logs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.187393 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-public-tls-certs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.187428 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-combined-ca-bundle\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.187480 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5tfs\" (UniqueName: \"kubernetes.io/projected/6907cde0-9a5e-432e-bc68-55b357a514d9-kube-api-access-q5tfs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.187511 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-internal-tls-certs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.289852 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5tfs\" (UniqueName: \"kubernetes.io/projected/6907cde0-9a5e-432e-bc68-55b357a514d9-kube-api-access-q5tfs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.289958 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-internal-tls-certs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.290009 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-config-data\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.290085 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-config-data-custom\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.290162 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6907cde0-9a5e-432e-bc68-55b357a514d9-logs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.290228 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-public-tls-certs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.290278 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-combined-ca-bundle\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.292622 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6907cde0-9a5e-432e-bc68-55b357a514d9-logs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.301647 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-combined-ca-bundle\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.304039 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-config-data-custom\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.315677 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-config-data\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.319013 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-internal-tls-certs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.324126 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5tfs\" (UniqueName: \"kubernetes.io/projected/6907cde0-9a5e-432e-bc68-55b357a514d9-kube-api-access-q5tfs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.324820 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6907cde0-9a5e-432e-bc68-55b357a514d9-public-tls-certs\") pod \"barbican-api-847b4dbb6-dsqmk\" (UID: \"6907cde0-9a5e-432e-bc68-55b357a514d9\") " pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: I1203 09:01:45.405213 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:45 crc kubenswrapper[4573]: W1203 09:01:45.458113 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90faa05a_6a95_4261_957f_44b45ea20a02.slice/crio-7d2a7956e549e3a393d94e1c0b347989be53fad2b22a1a0b0f792c9a9704ea69 WatchSource:0}: Error finding container 7d2a7956e549e3a393d94e1c0b347989be53fad2b22a1a0b0f792c9a9704ea69: Status 404 returned error can't find the container with id 7d2a7956e549e3a393d94e1c0b347989be53fad2b22a1a0b0f792c9a9704ea69 Dec 03 09:01:46 crc kubenswrapper[4573]: I1203 09:01:46.143983 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:46 crc kubenswrapper[4573]: I1203 09:01:46.157589 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-85bbcf664-5fm4n" Dec 03 09:01:46 crc kubenswrapper[4573]: I1203 09:01:46.212919 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerStarted","Data":"7d2a7956e549e3a393d94e1c0b347989be53fad2b22a1a0b0f792c9a9704ea69"} Dec 03 09:01:47 crc kubenswrapper[4573]: I1203 09:01:47.189253 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-847b4dbb6-dsqmk"] Dec 03 09:01:47 crc kubenswrapper[4573]: W1203 09:01:47.228240 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6907cde0_9a5e_432e_bc68_55b357a514d9.slice/crio-9ad7759dd68feb408afbb51068a58cf465b5581bbafdd3a11ebaf5ed4d00963e WatchSource:0}: Error finding container 9ad7759dd68feb408afbb51068a58cf465b5581bbafdd3a11ebaf5ed4d00963e: Status 404 returned error can't find the container with id 9ad7759dd68feb408afbb51068a58cf465b5581bbafdd3a11ebaf5ed4d00963e Dec 03 09:01:47 crc kubenswrapper[4573]: I1203 09:01:47.255867 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" event={"ID":"e954edc8-66c8-4f8f-a255-b8c5d810aa1a","Type":"ContainerStarted","Data":"a84b0bb6d903bd9fb82ed5bc2125bfa17e744027e3390f8d65f01318d80704f8"} Dec 03 09:01:47 crc kubenswrapper[4573]: I1203 09:01:47.291622 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-cd9d86d7-kn6nh" event={"ID":"865e87dd-1725-4932-9566-dbfbf7b85a60","Type":"ContainerStarted","Data":"a3a5c304217057cc998b7712613ef10d05f4d1fad1f93b2ea4237a954adad43f"} Dec 03 09:01:47 crc kubenswrapper[4573]: I1203 09:01:47.314605 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerStarted","Data":"06a981662731fec0c6a3b3b1df80b28b0c540b6de88e6c734f30a5524cee419a"} Dec 03 09:01:47 crc kubenswrapper[4573]: I1203 09:01:47.338925 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"04646290-c4ca-444e-b752-12a054221b16","Type":"ContainerStarted","Data":"7d5f1ed561de01d6577dc2b581f69c9716367522adaf3e078cdb85d67fd0a01c"} Dec 03 09:01:47 crc kubenswrapper[4573]: I1203 09:01:47.339454 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api-log" containerID="cri-o://d86adc4513427e0feed45141efce0fd0fb33dafcb7ec082b101bd0c0e3acd93d" gracePeriod=30 Dec 03 09:01:47 crc kubenswrapper[4573]: I1203 09:01:47.339892 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 09:01:47 crc kubenswrapper[4573]: I1203 09:01:47.340365 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api" containerID="cri-o://7d5f1ed561de01d6577dc2b581f69c9716367522adaf3e078cdb85d67fd0a01c" gracePeriod=30 Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.393004 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8c1c9237-25b7-41d3-bfc6-ac8dea228059","Type":"ContainerStarted","Data":"23b36c940df995323a22cf336b7f865b9e2ed33625eec92e2113f081635ec543"} Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.415955 4573 generic.go:334] "Generic (PLEG): container finished" podID="04646290-c4ca-444e-b752-12a054221b16" containerID="d86adc4513427e0feed45141efce0fd0fb33dafcb7ec082b101bd0c0e3acd93d" exitCode=143 Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.416686 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"04646290-c4ca-444e-b752-12a054221b16","Type":"ContainerDied","Data":"d86adc4513427e0feed45141efce0fd0fb33dafcb7ec082b101bd0c0e3acd93d"} Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.441396 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=8.587661977 podStartE2EDuration="10.441312419s" podCreationTimestamp="2025-12-03 09:01:38 +0000 UTC" firstStartedPulling="2025-12-03 09:01:41.025240732 +0000 UTC m=+1421.593619991" lastFinishedPulling="2025-12-03 09:01:42.878891174 +0000 UTC m=+1423.447270433" observedRunningTime="2025-12-03 09:01:48.426325537 +0000 UTC m=+1428.994704796" watchObservedRunningTime="2025-12-03 09:01:48.441312419 +0000 UTC m=+1429.009691678" Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.441602 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=10.441578057 podStartE2EDuration="10.441578057s" podCreationTimestamp="2025-12-03 09:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:47.391811442 +0000 UTC m=+1427.960190711" watchObservedRunningTime="2025-12-03 09:01:48.441578057 +0000 UTC m=+1429.009957316" Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.443311 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" event={"ID":"e954edc8-66c8-4f8f-a255-b8c5d810aa1a","Type":"ContainerStarted","Data":"d0bab31c3cd7deaa9a105cddde258cf30ba425a04edcc47488d8f77051f312db"} Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.463676 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847b4dbb6-dsqmk" event={"ID":"6907cde0-9a5e-432e-bc68-55b357a514d9","Type":"ContainerStarted","Data":"9e7f566f75397e8c06daa9da6dbf0e3aae81be9b38e60a2d7e41d1791c6426e9"} Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.463735 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847b4dbb6-dsqmk" event={"ID":"6907cde0-9a5e-432e-bc68-55b357a514d9","Type":"ContainerStarted","Data":"9ad7759dd68feb408afbb51068a58cf465b5581bbafdd3a11ebaf5ed4d00963e"} Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.506499 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5c965d87c4-dmtfc" podStartSLOduration=8.405315904 podStartE2EDuration="13.506466763s" podCreationTimestamp="2025-12-03 09:01:35 +0000 UTC" firstStartedPulling="2025-12-03 09:01:41.317419264 +0000 UTC m=+1421.885798523" lastFinishedPulling="2025-12-03 09:01:46.418570123 +0000 UTC m=+1426.986949382" observedRunningTime="2025-12-03 09:01:48.47764926 +0000 UTC m=+1429.046028519" watchObservedRunningTime="2025-12-03 09:01:48.506466763 +0000 UTC m=+1429.074846022" Dec 03 09:01:48 crc kubenswrapper[4573]: I1203 09:01:48.690029 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.092444 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.232377 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-v87zn"] Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.254767 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" podUID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" containerName="dnsmasq-dns" containerID="cri-o://71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd" gracePeriod=10 Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.537875 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-847b4dbb6-dsqmk" event={"ID":"6907cde0-9a5e-432e-bc68-55b357a514d9","Type":"ContainerStarted","Data":"5eb6dda1b32d021b1b2136c964ee431ac096f120ee20d6d821813ead9078b9a1"} Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.539651 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.539694 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.580300 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-cd9d86d7-kn6nh" event={"ID":"865e87dd-1725-4932-9566-dbfbf7b85a60","Type":"ContainerStarted","Data":"1fdbc44ca5f69850eab793819d59eee773543289178306f41c5cc40079ab5000"} Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.647460 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-847b4dbb6-dsqmk" podStartSLOduration=5.647441259 podStartE2EDuration="5.647441259s" podCreationTimestamp="2025-12-03 09:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:01:49.645921626 +0000 UTC m=+1430.214300895" watchObservedRunningTime="2025-12-03 09:01:49.647441259 +0000 UTC m=+1430.215820518" Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.649207 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerStarted","Data":"0297b7b023135e13bf67973ef377fc64ac67fc725226af9ac25fc6617e699a01"} Dec 03 09:01:49 crc kubenswrapper[4573]: I1203 09:01:49.704257 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-cd9d86d7-kn6nh" podStartSLOduration=8.986166062 podStartE2EDuration="14.704231761s" podCreationTimestamp="2025-12-03 09:01:35 +0000 UTC" firstStartedPulling="2025-12-03 09:01:40.699207029 +0000 UTC m=+1421.267586278" lastFinishedPulling="2025-12-03 09:01:46.417272718 +0000 UTC m=+1426.985651977" observedRunningTime="2025-12-03 09:01:49.702341059 +0000 UTC m=+1430.270720318" watchObservedRunningTime="2025-12-03 09:01:49.704231761 +0000 UTC m=+1430.272611020" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.301275 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.410265 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdftc\" (UniqueName: \"kubernetes.io/projected/317e24d7-2fe9-412b-8fb7-1ccb697b12de-kube-api-access-pdftc\") pod \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.410371 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-nb\") pod \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.410427 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-sb\") pod \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.410572 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-config\") pod \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.410612 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-swift-storage-0\") pod \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.410665 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-svc\") pod \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\" (UID: \"317e24d7-2fe9-412b-8fb7-1ccb697b12de\") " Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.440523 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/317e24d7-2fe9-412b-8fb7-1ccb697b12de-kube-api-access-pdftc" (OuterVolumeSpecName: "kube-api-access-pdftc") pod "317e24d7-2fe9-412b-8fb7-1ccb697b12de" (UID: "317e24d7-2fe9-412b-8fb7-1ccb697b12de"). InnerVolumeSpecName "kube-api-access-pdftc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.515638 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdftc\" (UniqueName: \"kubernetes.io/projected/317e24d7-2fe9-412b-8fb7-1ccb697b12de-kube-api-access-pdftc\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.540290 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "317e24d7-2fe9-412b-8fb7-1ccb697b12de" (UID: "317e24d7-2fe9-412b-8fb7-1ccb697b12de"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.604708 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "317e24d7-2fe9-412b-8fb7-1ccb697b12de" (UID: "317e24d7-2fe9-412b-8fb7-1ccb697b12de"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.607035 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-config" (OuterVolumeSpecName: "config") pod "317e24d7-2fe9-412b-8fb7-1ccb697b12de" (UID: "317e24d7-2fe9-412b-8fb7-1ccb697b12de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.619368 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.619416 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.619429 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.639573 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "317e24d7-2fe9-412b-8fb7-1ccb697b12de" (UID: "317e24d7-2fe9-412b-8fb7-1ccb697b12de"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.642578 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "317e24d7-2fe9-412b-8fb7-1ccb697b12de" (UID: "317e24d7-2fe9-412b-8fb7-1ccb697b12de"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.695589 4573 generic.go:334] "Generic (PLEG): container finished" podID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" containerID="71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd" exitCode=0 Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.695753 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" event={"ID":"317e24d7-2fe9-412b-8fb7-1ccb697b12de","Type":"ContainerDied","Data":"71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd"} Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.695795 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" event={"ID":"317e24d7-2fe9-412b-8fb7-1ccb697b12de","Type":"ContainerDied","Data":"30e65a70ace00f0cd9c6f570526aaced5a203cc0a3d8dcb4a300b90b0d00b509"} Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.695816 4573 scope.go:117] "RemoveContainer" containerID="71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.696028 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-v87zn" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.717770 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerStarted","Data":"cac51223d774810a3e39697796598123271ab044463433d6e9d852054fe47f66"} Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.724411 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.724759 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/317e24d7-2fe9-412b-8fb7-1ccb697b12de-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.740202 4573 scope.go:117] "RemoveContainer" containerID="5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.799003 4573 scope.go:117] "RemoveContainer" containerID="71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd" Dec 03 09:01:50 crc kubenswrapper[4573]: E1203 09:01:50.812033 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd\": container with ID starting with 71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd not found: ID does not exist" containerID="71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.812122 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd"} err="failed to get container status \"71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd\": rpc error: code = NotFound desc = could not find container \"71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd\": container with ID starting with 71dc068bf9d44076f81308b9f7acf59b8f006f48faca259d7ce67975849533cd not found: ID does not exist" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.812157 4573 scope.go:117] "RemoveContainer" containerID="5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629" Dec 03 09:01:50 crc kubenswrapper[4573]: E1203 09:01:50.820237 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629\": container with ID starting with 5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629 not found: ID does not exist" containerID="5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.820289 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629"} err="failed to get container status \"5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629\": rpc error: code = NotFound desc = could not find container \"5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629\": container with ID starting with 5b2f5ed7033d35fc6eca8b998c31b36e52673a8ad8cb43339c587ec3d0514629 not found: ID does not exist" Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.839562 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-v87zn"] Dec 03 09:01:50 crc kubenswrapper[4573]: I1203 09:01:50.866631 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-v87zn"] Dec 03 09:01:51 crc kubenswrapper[4573]: I1203 09:01:51.772027 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:52 crc kubenswrapper[4573]: I1203 09:01:52.049819 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" path="/var/lib/kubelet/pods/317e24d7-2fe9-412b-8fb7-1ccb697b12de/volumes" Dec 03 09:01:52 crc kubenswrapper[4573]: I1203 09:01:52.757710 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerStarted","Data":"434e113044392686995debedfeaeba680f9eb5c738e1a1c4235fe23fda564b20"} Dec 03 09:01:52 crc kubenswrapper[4573]: I1203 09:01:52.759439 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:01:52 crc kubenswrapper[4573]: I1203 09:01:52.803801 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.6276831309999995 podStartE2EDuration="11.803782027s" podCreationTimestamp="2025-12-03 09:01:41 +0000 UTC" firstStartedPulling="2025-12-03 09:01:45.481577203 +0000 UTC m=+1426.049956462" lastFinishedPulling="2025-12-03 09:01:51.657676099 +0000 UTC m=+1432.226055358" observedRunningTime="2025-12-03 09:01:52.801068622 +0000 UTC m=+1433.369447891" watchObservedRunningTime="2025-12-03 09:01:52.803782027 +0000 UTC m=+1433.372161286" Dec 03 09:01:53 crc kubenswrapper[4573]: I1203 09:01:53.811332 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:53 crc kubenswrapper[4573]: I1203 09:01:53.811431 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:54 crc kubenswrapper[4573]: I1203 09:01:54.762684 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 09:01:54 crc kubenswrapper[4573]: I1203 09:01:54.814663 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:01:54 crc kubenswrapper[4573]: I1203 09:01:54.814932 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerName="cinder-scheduler" containerID="cri-o://35675082d40a9c2fd01697ae0adc20fc68defb5adb12ce6f6ea13227526d79c5" gracePeriod=30 Dec 03 09:01:54 crc kubenswrapper[4573]: I1203 09:01:54.817021 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerName="probe" containerID="cri-o://23b36c940df995323a22cf336b7f865b9e2ed33625eec92e2113f081635ec543" gracePeriod=30 Dec 03 09:01:54 crc kubenswrapper[4573]: I1203 09:01:54.830518 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-f56bd7d55-5qgxb" Dec 03 09:01:55 crc kubenswrapper[4573]: I1203 09:01:55.524266 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.151163 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.152197 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.152238 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.259987 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.994281 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 09:01:57 crc kubenswrapper[4573]: E1203 09:01:57.995339 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" containerName="dnsmasq-dns" Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.995361 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" containerName="dnsmasq-dns" Dec 03 09:01:57 crc kubenswrapper[4573]: E1203 09:01:57.995410 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" containerName="init" Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.995416 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" containerName="init" Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.995672 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="317e24d7-2fe9-412b-8fb7-1ccb697b12de" containerName="dnsmasq-dns" Dec 03 09:01:57 crc kubenswrapper[4573]: I1203 09:01:57.997596 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.003693 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.005483 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jxd7k" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.005719 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.024534 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.122030 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz4nn\" (UniqueName: \"kubernetes.io/projected/4357c5a1-7325-4bff-8a79-4ffd70879b1c-kube-api-access-dz4nn\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.122108 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4357c5a1-7325-4bff-8a79-4ffd70879b1c-openstack-config-secret\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.122136 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4357c5a1-7325-4bff-8a79-4ffd70879b1c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.122273 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4357c5a1-7325-4bff-8a79-4ffd70879b1c-openstack-config\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.173486 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerID="23b36c940df995323a22cf336b7f865b9e2ed33625eec92e2113f081635ec543" exitCode=0 Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.173965 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerID="35675082d40a9c2fd01697ae0adc20fc68defb5adb12ce6f6ea13227526d79c5" exitCode=0 Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.173994 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8c1c9237-25b7-41d3-bfc6-ac8dea228059","Type":"ContainerDied","Data":"23b36c940df995323a22cf336b7f865b9e2ed33625eec92e2113f081635ec543"} Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.174036 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8c1c9237-25b7-41d3-bfc6-ac8dea228059","Type":"ContainerDied","Data":"35675082d40a9c2fd01697ae0adc20fc68defb5adb12ce6f6ea13227526d79c5"} Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.226637 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4357c5a1-7325-4bff-8a79-4ffd70879b1c-openstack-config\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.226776 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz4nn\" (UniqueName: \"kubernetes.io/projected/4357c5a1-7325-4bff-8a79-4ffd70879b1c-kube-api-access-dz4nn\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.226803 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4357c5a1-7325-4bff-8a79-4ffd70879b1c-openstack-config-secret\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.226825 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4357c5a1-7325-4bff-8a79-4ffd70879b1c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.230496 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4357c5a1-7325-4bff-8a79-4ffd70879b1c-openstack-config\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.236484 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4357c5a1-7325-4bff-8a79-4ffd70879b1c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.253019 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4357c5a1-7325-4bff-8a79-4ffd70879b1c-openstack-config-secret\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.271960 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz4nn\" (UniqueName: \"kubernetes.io/projected/4357c5a1-7325-4bff-8a79-4ffd70879b1c-kube-api-access-dz4nn\") pod \"openstackclient\" (UID: \"4357c5a1-7325-4bff-8a79-4ffd70879b1c\") " pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.330271 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.918454 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:58 crc kubenswrapper[4573]: I1203 09:01:58.919309 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.045376 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.161428 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data\") pod \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.161497 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-combined-ca-bundle\") pod \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.161586 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8fpc\" (UniqueName: \"kubernetes.io/projected/8c1c9237-25b7-41d3-bfc6-ac8dea228059-kube-api-access-l8fpc\") pod \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.161792 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c1c9237-25b7-41d3-bfc6-ac8dea228059-etc-machine-id\") pod \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.161952 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-scripts\") pod \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.161987 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data-custom\") pod \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\" (UID: \"8c1c9237-25b7-41d3-bfc6-ac8dea228059\") " Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.166965 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c1c9237-25b7-41d3-bfc6-ac8dea228059-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8c1c9237-25b7-41d3-bfc6-ac8dea228059" (UID: "8c1c9237-25b7-41d3-bfc6-ac8dea228059"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.182232 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-scripts" (OuterVolumeSpecName: "scripts") pod "8c1c9237-25b7-41d3-bfc6-ac8dea228059" (UID: "8c1c9237-25b7-41d3-bfc6-ac8dea228059"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.192625 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c1c9237-25b7-41d3-bfc6-ac8dea228059-kube-api-access-l8fpc" (OuterVolumeSpecName: "kube-api-access-l8fpc") pod "8c1c9237-25b7-41d3-bfc6-ac8dea228059" (UID: "8c1c9237-25b7-41d3-bfc6-ac8dea228059"). InnerVolumeSpecName "kube-api-access-l8fpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.193845 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8c1c9237-25b7-41d3-bfc6-ac8dea228059" (UID: "8c1c9237-25b7-41d3-bfc6-ac8dea228059"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.234362 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8c1c9237-25b7-41d3-bfc6-ac8dea228059","Type":"ContainerDied","Data":"7418dab3f76d21a40e83486cfb1b258ccb71ab1e5b39d13b52eeeeb540e79d3c"} Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.234447 4573 scope.go:117] "RemoveContainer" containerID="23b36c940df995323a22cf336b7f865b9e2ed33625eec92e2113f081635ec543" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.234628 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.263347 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.163:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.268145 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8fpc\" (UniqueName: \"kubernetes.io/projected/8c1c9237-25b7-41d3-bfc6-ac8dea228059-kube-api-access-l8fpc\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.268412 4573 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c1c9237-25b7-41d3-bfc6-ac8dea228059-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.268493 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.268568 4573 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.347015 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.354369 4573 scope.go:117] "RemoveContainer" containerID="35675082d40a9c2fd01697ae0adc20fc68defb5adb12ce6f6ea13227526d79c5" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.451496 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-847b4dbb6-dsqmk" podUID="6907cde0-9a5e-432e-bc68-55b357a514d9" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.452096 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-847b4dbb6-dsqmk" podUID="6907cde0-9a5e-432e-bc68-55b357a514d9" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.486616 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c1c9237-25b7-41d3-bfc6-ac8dea228059" (UID: "8c1c9237-25b7-41d3-bfc6-ac8dea228059"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.526778 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data" (OuterVolumeSpecName: "config-data") pod "8c1c9237-25b7-41d3-bfc6-ac8dea228059" (UID: "8c1c9237-25b7-41d3-bfc6-ac8dea228059"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.583933 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.583982 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c1c9237-25b7-41d3-bfc6-ac8dea228059-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.644200 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.670472 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.696407 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:01:59 crc kubenswrapper[4573]: E1203 09:01:59.697390 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerName="cinder-scheduler" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.697413 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerName="cinder-scheduler" Dec 03 09:01:59 crc kubenswrapper[4573]: E1203 09:01:59.697437 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerName="probe" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.697444 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerName="probe" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.697652 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerName="probe" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.697671 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" containerName="cinder-scheduler" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.699132 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.707310 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.720649 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.790754 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-config-data\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.790827 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20097614-b580-49bf-a8c7-04f371825b9a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.790877 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-scripts\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.790934 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.790964 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.790998 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmmkp\" (UniqueName: \"kubernetes.io/projected/20097614-b580-49bf-a8c7-04f371825b9a-kube-api-access-xmmkp\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.892873 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-scripts\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.892960 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.892997 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.893031 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmmkp\" (UniqueName: \"kubernetes.io/projected/20097614-b580-49bf-a8c7-04f371825b9a-kube-api-access-xmmkp\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.893170 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-config-data\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.893202 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20097614-b580-49bf-a8c7-04f371825b9a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.893300 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20097614-b580-49bf-a8c7-04f371825b9a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.904689 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.905688 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-scripts\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.909982 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-config-data\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.923866 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmmkp\" (UniqueName: \"kubernetes.io/projected/20097614-b580-49bf-a8c7-04f371825b9a-kube-api-access-xmmkp\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:01:59 crc kubenswrapper[4573]: I1203 09:01:59.925978 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20097614-b580-49bf-a8c7-04f371825b9a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"20097614-b580-49bf-a8c7-04f371825b9a\") " pod="openstack/cinder-scheduler-0" Dec 03 09:02:00 crc kubenswrapper[4573]: I1203 09:02:00.055556 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 09:02:00 crc kubenswrapper[4573]: I1203 09:02:00.123900 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c1c9237-25b7-41d3-bfc6-ac8dea228059" path="/var/lib/kubelet/pods/8c1c9237-25b7-41d3-bfc6-ac8dea228059/volumes" Dec 03 09:02:00 crc kubenswrapper[4573]: I1203 09:02:00.313717 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4357c5a1-7325-4bff-8a79-4ffd70879b1c","Type":"ContainerStarted","Data":"58b0739848327bcce0055a63cabf7beb0be68b6b7a13498b101af86357fe2833"} Dec 03 09:02:00 crc kubenswrapper[4573]: I1203 09:02:00.417198 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-847b4dbb6-dsqmk" podUID="6907cde0-9a5e-432e-bc68-55b357a514d9" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:00 crc kubenswrapper[4573]: I1203 09:02:00.417574 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-847b4dbb6-dsqmk" podUID="6907cde0-9a5e-432e-bc68-55b357a514d9" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:01 crc kubenswrapper[4573]: I1203 09:02:01.038407 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 09:02:01 crc kubenswrapper[4573]: I1203 09:02:01.353302 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20097614-b580-49bf-a8c7-04f371825b9a","Type":"ContainerStarted","Data":"9c8c4a95f925dc86d30310023278b8b33db42650a1039451eccfe766e54d6136"} Dec 03 09:02:01 crc kubenswrapper[4573]: I1203 09:02:01.361318 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerID="875e1f86deecceeb9c350e1909663db37b4f6f0a82cff5c1165cad17d9cdc1df" exitCode=137 Dec 03 09:02:01 crc kubenswrapper[4573]: I1203 09:02:01.361416 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerDied","Data":"875e1f86deecceeb9c350e1909663db37b4f6f0a82cff5c1165cad17d9cdc1df"} Dec 03 09:02:02 crc kubenswrapper[4573]: I1203 09:02:02.162273 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:02 crc kubenswrapper[4573]: I1203 09:02:02.413010 4573 generic.go:334] "Generic (PLEG): container finished" podID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerID="afd0f349dd0be138a212ad8426fb19507ce3fcb56dbe6dbd953d89456cd91458" exitCode=137 Dec 03 09:02:02 crc kubenswrapper[4573]: I1203 09:02:02.413138 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66f99c579b-gh7sh" event={"ID":"b78e2453-adc3-4139-97c5-9f8a977df1ee","Type":"ContainerDied","Data":"afd0f349dd0be138a212ad8426fb19507ce3fcb56dbe6dbd953d89456cd91458"} Dec 03 09:02:02 crc kubenswrapper[4573]: I1203 09:02:02.419267 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerStarted","Data":"b8aa7ba7c95bd0a8692253b9b93fecbaa60bed28d93ea2c5355468e1a8ef3047"} Dec 03 09:02:03 crc kubenswrapper[4573]: I1203 09:02:03.456499 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66f99c579b-gh7sh" event={"ID":"b78e2453-adc3-4139-97c5-9f8a977df1ee","Type":"ContainerStarted","Data":"eee5245aaff562e364c268a895dd3a53869d1cf3cec162d72a6fbfa629898941"} Dec 03 09:02:03 crc kubenswrapper[4573]: I1203 09:02:03.486477 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20097614-b580-49bf-a8c7-04f371825b9a","Type":"ContainerStarted","Data":"03dd1a932d9b996cc878520d116899a06f12f362a8c01a53588b57a436f3d3d2"} Dec 03 09:02:04 crc kubenswrapper[4573]: I1203 09:02:04.307376 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.163:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:04 crc kubenswrapper[4573]: I1203 09:02:04.452437 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-847b4dbb6-dsqmk" podUID="6907cde0-9a5e-432e-bc68-55b357a514d9" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:04 crc kubenswrapper[4573]: I1203 09:02:04.507578 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"20097614-b580-49bf-a8c7-04f371825b9a","Type":"ContainerStarted","Data":"ce600ba99694ccc8bd2d4891483b66b037d3c6239b98c28660d5e9c6b597e930"} Dec 03 09:02:04 crc kubenswrapper[4573]: I1203 09:02:04.536403 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.536384296 podStartE2EDuration="5.536384296s" podCreationTimestamp="2025-12-03 09:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:02:04.53255714 +0000 UTC m=+1445.100936399" watchObservedRunningTime="2025-12-03 09:02:04.536384296 +0000 UTC m=+1445.104763555" Dec 03 09:02:05 crc kubenswrapper[4573]: I1203 09:02:05.056326 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 09:02:05 crc kubenswrapper[4573]: I1203 09:02:05.464367 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-847b4dbb6-dsqmk" podUID="6907cde0-9a5e-432e-bc68-55b357a514d9" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:05 crc kubenswrapper[4573]: I1203 09:02:05.464367 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-847b4dbb6-dsqmk" podUID="6907cde0-9a5e-432e-bc68-55b357a514d9" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.165:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:05 crc kubenswrapper[4573]: I1203 09:02:05.499475 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:02:05 crc kubenswrapper[4573]: I1203 09:02:05.500982 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-847b4dbb6-dsqmk" Dec 03 09:02:05 crc kubenswrapper[4573]: I1203 09:02:05.689881 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6fbfd76774-v2hm7"] Dec 03 09:02:05 crc kubenswrapper[4573]: I1203 09:02:05.691798 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" containerID="cri-o://3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86" gracePeriod=30 Dec 03 09:02:05 crc kubenswrapper[4573]: I1203 09:02:05.692433 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api" containerID="cri-o://175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316" gracePeriod=30 Dec 03 09:02:06 crc kubenswrapper[4573]: I1203 09:02:06.547459 4573 generic.go:334] "Generic (PLEG): container finished" podID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerID="3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86" exitCode=143 Dec 03 09:02:06 crc kubenswrapper[4573]: I1203 09:02:06.548113 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fbfd76774-v2hm7" event={"ID":"a37c63f1-b984-4c85-9866-e54541dc09b8","Type":"ContainerDied","Data":"3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86"} Dec 03 09:02:09 crc kubenswrapper[4573]: I1203 09:02:09.349345 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.163:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:09 crc kubenswrapper[4573]: I1203 09:02:09.512107 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:38666->10.217.0.160:9311: read: connection reset by peer" Dec 03 09:02:09 crc kubenswrapper[4573]: I1203 09:02:09.512155 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6fbfd76774-v2hm7" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:38676->10.217.0.160:9311: read: connection reset by peer" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.169209 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.309143 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-combined-ca-bundle\") pod \"a37c63f1-b984-4c85-9866-e54541dc09b8\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.309232 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz48z\" (UniqueName: \"kubernetes.io/projected/a37c63f1-b984-4c85-9866-e54541dc09b8-kube-api-access-kz48z\") pod \"a37c63f1-b984-4c85-9866-e54541dc09b8\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.309367 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data\") pod \"a37c63f1-b984-4c85-9866-e54541dc09b8\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.309573 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a37c63f1-b984-4c85-9866-e54541dc09b8-logs\") pod \"a37c63f1-b984-4c85-9866-e54541dc09b8\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.309626 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data-custom\") pod \"a37c63f1-b984-4c85-9866-e54541dc09b8\" (UID: \"a37c63f1-b984-4c85-9866-e54541dc09b8\") " Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.310411 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a37c63f1-b984-4c85-9866-e54541dc09b8-logs" (OuterVolumeSpecName: "logs") pod "a37c63f1-b984-4c85-9866-e54541dc09b8" (UID: "a37c63f1-b984-4c85-9866-e54541dc09b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.310684 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a37c63f1-b984-4c85-9866-e54541dc09b8-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.320256 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a37c63f1-b984-4c85-9866-e54541dc09b8" (UID: "a37c63f1-b984-4c85-9866-e54541dc09b8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.366079 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a37c63f1-b984-4c85-9866-e54541dc09b8-kube-api-access-kz48z" (OuterVolumeSpecName: "kube-api-access-kz48z") pod "a37c63f1-b984-4c85-9866-e54541dc09b8" (UID: "a37c63f1-b984-4c85-9866-e54541dc09b8"). InnerVolumeSpecName "kube-api-access-kz48z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.413420 4573 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.413461 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz48z\" (UniqueName: \"kubernetes.io/projected/a37c63f1-b984-4c85-9866-e54541dc09b8-kube-api-access-kz48z\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.417254 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a37c63f1-b984-4c85-9866-e54541dc09b8" (UID: "a37c63f1-b984-4c85-9866-e54541dc09b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.461299 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data" (OuterVolumeSpecName: "config-data") pod "a37c63f1-b984-4c85-9866-e54541dc09b8" (UID: "a37c63f1-b984-4c85-9866-e54541dc09b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.483021 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.516415 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.516461 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a37c63f1-b984-4c85-9866-e54541dc09b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.601719 4573 generic.go:334] "Generic (PLEG): container finished" podID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerID="175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316" exitCode=0 Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.601788 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fbfd76774-v2hm7" event={"ID":"a37c63f1-b984-4c85-9866-e54541dc09b8","Type":"ContainerDied","Data":"175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316"} Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.601837 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fbfd76774-v2hm7" event={"ID":"a37c63f1-b984-4c85-9866-e54541dc09b8","Type":"ContainerDied","Data":"6326635287a540d0d0740b2fdf3c8534aee273f1953626ae766cf7eb1158c091"} Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.601865 4573 scope.go:117] "RemoveContainer" containerID="175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.601884 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fbfd76774-v2hm7" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.696668 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6fbfd76774-v2hm7"] Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.710545 4573 scope.go:117] "RemoveContainer" containerID="3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.710907 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6fbfd76774-v2hm7"] Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.752293 4573 scope.go:117] "RemoveContainer" containerID="175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316" Dec 03 09:02:10 crc kubenswrapper[4573]: E1203 09:02:10.753117 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316\": container with ID starting with 175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316 not found: ID does not exist" containerID="175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.753219 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316"} err="failed to get container status \"175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316\": rpc error: code = NotFound desc = could not find container \"175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316\": container with ID starting with 175ed4693146e672e2daeeb448d97ad11703b5d0dee5b47f2166d6151725f316 not found: ID does not exist" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.753269 4573 scope.go:117] "RemoveContainer" containerID="3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86" Dec 03 09:02:10 crc kubenswrapper[4573]: E1203 09:02:10.753867 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86\": container with ID starting with 3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86 not found: ID does not exist" containerID="3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.753893 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86"} err="failed to get container status \"3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86\": rpc error: code = NotFound desc = could not find container \"3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86\": container with ID starting with 3c524757c3fdf9a56f7242156ca01d8165a1bff374d38d2431a260763a886e86 not found: ID does not exist" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.918920 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:02:10 crc kubenswrapper[4573]: I1203 09:02:10.919866 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:02:11 crc kubenswrapper[4573]: I1203 09:02:11.070606 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:02:11 crc kubenswrapper[4573]: I1203 09:02:11.071582 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:02:12 crc kubenswrapper[4573]: I1203 09:02:12.044653 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" path="/var/lib/kubelet/pods/a37c63f1-b984-4c85-9866-e54541dc09b8/volumes" Dec 03 09:02:12 crc kubenswrapper[4573]: I1203 09:02:12.360555 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 09:02:12 crc kubenswrapper[4573]: I1203 09:02:12.902799 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.353016 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-66d678ddfc-gvbk5"] Dec 03 09:02:13 crc kubenswrapper[4573]: E1203 09:02:13.353680 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.353706 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" Dec 03 09:02:13 crc kubenswrapper[4573]: E1203 09:02:13.353757 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.353767 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.354014 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api-log" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.354096 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="a37c63f1-b984-4c85-9866-e54541dc09b8" containerName="barbican-api" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.355624 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.359758 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.359991 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.363243 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.400550 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-66d678ddfc-gvbk5"] Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.495110 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-log-httpd\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.495308 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-run-httpd\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.495508 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-combined-ca-bundle\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.495669 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-public-tls-certs\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.495702 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29dz5\" (UniqueName: \"kubernetes.io/projected/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-kube-api-access-29dz5\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.495816 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-internal-tls-certs\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.495862 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-config-data\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.495890 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-etc-swift\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.598709 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-log-httpd\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.598779 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-run-httpd\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.598835 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-combined-ca-bundle\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.598885 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-public-tls-certs\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.598906 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29dz5\" (UniqueName: \"kubernetes.io/projected/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-kube-api-access-29dz5\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.598947 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-internal-tls-certs\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.598968 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-config-data\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.598984 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-etc-swift\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.601734 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-run-httpd\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.602136 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-log-httpd\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.610145 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-etc-swift\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.611697 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-public-tls-certs\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.616975 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-combined-ca-bundle\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.618337 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-config-data\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.626884 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29dz5\" (UniqueName: \"kubernetes.io/projected/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-kube-api-access-29dz5\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.629232 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/02bc648a-d9c8-4fdf-bf48-bc57a6b05386-internal-tls-certs\") pod \"swift-proxy-66d678ddfc-gvbk5\" (UID: \"02bc648a-d9c8-4fdf-bf48-bc57a6b05386\") " pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:13 crc kubenswrapper[4573]: I1203 09:02:13.694166 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:15 crc kubenswrapper[4573]: I1203 09:02:15.229869 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:15 crc kubenswrapper[4573]: I1203 09:02:15.230291 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="ceilometer-central-agent" containerID="cri-o://06a981662731fec0c6a3b3b1df80b28b0c540b6de88e6c734f30a5524cee419a" gracePeriod=30 Dec 03 09:02:15 crc kubenswrapper[4573]: I1203 09:02:15.230371 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="proxy-httpd" containerID="cri-o://434e113044392686995debedfeaeba680f9eb5c738e1a1c4235fe23fda564b20" gracePeriod=30 Dec 03 09:02:15 crc kubenswrapper[4573]: I1203 09:02:15.230457 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="ceilometer-notification-agent" containerID="cri-o://0297b7b023135e13bf67973ef377fc64ac67fc725226af9ac25fc6617e699a01" gracePeriod=30 Dec 03 09:02:15 crc kubenswrapper[4573]: I1203 09:02:15.230464 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="sg-core" containerID="cri-o://cac51223d774810a3e39697796598123271ab044463433d6e9d852054fe47f66" gracePeriod=30 Dec 03 09:02:16 crc kubenswrapper[4573]: I1203 09:02:16.711834 4573 generic.go:334] "Generic (PLEG): container finished" podID="90faa05a-6a95-4261-957f-44b45ea20a02" containerID="434e113044392686995debedfeaeba680f9eb5c738e1a1c4235fe23fda564b20" exitCode=0 Dec 03 09:02:16 crc kubenswrapper[4573]: I1203 09:02:16.712308 4573 generic.go:334] "Generic (PLEG): container finished" podID="90faa05a-6a95-4261-957f-44b45ea20a02" containerID="cac51223d774810a3e39697796598123271ab044463433d6e9d852054fe47f66" exitCode=2 Dec 03 09:02:16 crc kubenswrapper[4573]: I1203 09:02:16.712318 4573 generic.go:334] "Generic (PLEG): container finished" podID="90faa05a-6a95-4261-957f-44b45ea20a02" containerID="0297b7b023135e13bf67973ef377fc64ac67fc725226af9ac25fc6617e699a01" exitCode=0 Dec 03 09:02:16 crc kubenswrapper[4573]: I1203 09:02:16.712330 4573 generic.go:334] "Generic (PLEG): container finished" podID="90faa05a-6a95-4261-957f-44b45ea20a02" containerID="06a981662731fec0c6a3b3b1df80b28b0c540b6de88e6c734f30a5524cee419a" exitCode=0 Dec 03 09:02:16 crc kubenswrapper[4573]: I1203 09:02:16.711921 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerDied","Data":"434e113044392686995debedfeaeba680f9eb5c738e1a1c4235fe23fda564b20"} Dec 03 09:02:16 crc kubenswrapper[4573]: I1203 09:02:16.712370 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerDied","Data":"cac51223d774810a3e39697796598123271ab044463433d6e9d852054fe47f66"} Dec 03 09:02:16 crc kubenswrapper[4573]: I1203 09:02:16.712384 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerDied","Data":"0297b7b023135e13bf67973ef377fc64ac67fc725226af9ac25fc6617e699a01"} Dec 03 09:02:16 crc kubenswrapper[4573]: I1203 09:02:16.712395 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerDied","Data":"06a981662731fec0c6a3b3b1df80b28b0c540b6de88e6c734f30a5524cee419a"} Dec 03 09:02:18 crc kubenswrapper[4573]: I1203 09:02:18.753803 4573 generic.go:334] "Generic (PLEG): container finished" podID="04646290-c4ca-444e-b752-12a054221b16" containerID="7d5f1ed561de01d6577dc2b581f69c9716367522adaf3e078cdb85d67fd0a01c" exitCode=137 Dec 03 09:02:18 crc kubenswrapper[4573]: I1203 09:02:18.754237 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"04646290-c4ca-444e-b752-12a054221b16","Type":"ContainerDied","Data":"7d5f1ed561de01d6577dc2b581f69c9716367522adaf3e078cdb85d67fd0a01c"} Dec 03 09:02:19 crc kubenswrapper[4573]: I1203 09:02:19.203149 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.163:8776/healthcheck\": dial tcp 10.217.0.163:8776: connect: connection refused" Dec 03 09:02:20 crc kubenswrapper[4573]: I1203 09:02:20.920874 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 09:02:21 crc kubenswrapper[4573]: I1203 09:02:21.072694 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.691643 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.778281 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data\") pod \"04646290-c4ca-444e-b752-12a054221b16\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.778400 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04646290-c4ca-444e-b752-12a054221b16-logs\") pod \"04646290-c4ca-444e-b752-12a054221b16\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.778442 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data-custom\") pod \"04646290-c4ca-444e-b752-12a054221b16\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.778483 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6sll\" (UniqueName: \"kubernetes.io/projected/04646290-c4ca-444e-b752-12a054221b16-kube-api-access-g6sll\") pod \"04646290-c4ca-444e-b752-12a054221b16\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.778538 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04646290-c4ca-444e-b752-12a054221b16-etc-machine-id\") pod \"04646290-c4ca-444e-b752-12a054221b16\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.778689 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-combined-ca-bundle\") pod \"04646290-c4ca-444e-b752-12a054221b16\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.778732 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-scripts\") pod \"04646290-c4ca-444e-b752-12a054221b16\" (UID: \"04646290-c4ca-444e-b752-12a054221b16\") " Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.787021 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04646290-c4ca-444e-b752-12a054221b16-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "04646290-c4ca-444e-b752-12a054221b16" (UID: "04646290-c4ca-444e-b752-12a054221b16"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.787809 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04646290-c4ca-444e-b752-12a054221b16-logs" (OuterVolumeSpecName: "logs") pod "04646290-c4ca-444e-b752-12a054221b16" (UID: "04646290-c4ca-444e-b752-12a054221b16"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.806706 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-scripts" (OuterVolumeSpecName: "scripts") pod "04646290-c4ca-444e-b752-12a054221b16" (UID: "04646290-c4ca-444e-b752-12a054221b16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.810225 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "04646290-c4ca-444e-b752-12a054221b16" (UID: "04646290-c4ca-444e-b752-12a054221b16"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.825509 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04646290-c4ca-444e-b752-12a054221b16-kube-api-access-g6sll" (OuterVolumeSpecName: "kube-api-access-g6sll") pod "04646290-c4ca-444e-b752-12a054221b16" (UID: "04646290-c4ca-444e-b752-12a054221b16"). InnerVolumeSpecName "kube-api-access-g6sll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.846560 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"04646290-c4ca-444e-b752-12a054221b16","Type":"ContainerDied","Data":"77e9185a104bdca0d50f65af81ba892a2ec93290cee0019eaddb3e0933bff284"} Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.846633 4573 scope.go:117] "RemoveContainer" containerID="7d5f1ed561de01d6577dc2b581f69c9716367522adaf3e078cdb85d67fd0a01c" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.846818 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.886805 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.886857 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04646290-c4ca-444e-b752-12a054221b16-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.886872 4573 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.886884 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6sll\" (UniqueName: \"kubernetes.io/projected/04646290-c4ca-444e-b752-12a054221b16-kube-api-access-g6sll\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.886912 4573 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/04646290-c4ca-444e-b752-12a054221b16-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.906142 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04646290-c4ca-444e-b752-12a054221b16" (UID: "04646290-c4ca-444e-b752-12a054221b16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.908182 4573 scope.go:117] "RemoveContainer" containerID="d86adc4513427e0feed45141efce0fd0fb33dafcb7ec082b101bd0c0e3acd93d" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.961286 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.985068 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data" (OuterVolumeSpecName: "config-data") pod "04646290-c4ca-444e-b752-12a054221b16" (UID: "04646290-c4ca-444e-b752-12a054221b16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.989205 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:22 crc kubenswrapper[4573]: I1203 09:02:22.989233 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04646290-c4ca-444e-b752-12a054221b16-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.097765 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-sg-core-conf-yaml\") pod \"90faa05a-6a95-4261-957f-44b45ea20a02\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.097889 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-config-data\") pod \"90faa05a-6a95-4261-957f-44b45ea20a02\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.097963 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-run-httpd\") pod \"90faa05a-6a95-4261-957f-44b45ea20a02\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.098018 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-combined-ca-bundle\") pod \"90faa05a-6a95-4261-957f-44b45ea20a02\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.098143 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-log-httpd\") pod \"90faa05a-6a95-4261-957f-44b45ea20a02\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.098183 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-scripts\") pod \"90faa05a-6a95-4261-957f-44b45ea20a02\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.098214 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5fnd\" (UniqueName: \"kubernetes.io/projected/90faa05a-6a95-4261-957f-44b45ea20a02-kube-api-access-l5fnd\") pod \"90faa05a-6a95-4261-957f-44b45ea20a02\" (UID: \"90faa05a-6a95-4261-957f-44b45ea20a02\") " Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.100646 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "90faa05a-6a95-4261-957f-44b45ea20a02" (UID: "90faa05a-6a95-4261-957f-44b45ea20a02"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.114179 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90faa05a-6a95-4261-957f-44b45ea20a02-kube-api-access-l5fnd" (OuterVolumeSpecName: "kube-api-access-l5fnd") pod "90faa05a-6a95-4261-957f-44b45ea20a02" (UID: "90faa05a-6a95-4261-957f-44b45ea20a02"). InnerVolumeSpecName "kube-api-access-l5fnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.114983 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "90faa05a-6a95-4261-957f-44b45ea20a02" (UID: "90faa05a-6a95-4261-957f-44b45ea20a02"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.122113 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-scripts" (OuterVolumeSpecName: "scripts") pod "90faa05a-6a95-4261-957f-44b45ea20a02" (UID: "90faa05a-6a95-4261-957f-44b45ea20a02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.201359 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.201415 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/90faa05a-6a95-4261-957f-44b45ea20a02-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.201429 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.201440 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5fnd\" (UniqueName: \"kubernetes.io/projected/90faa05a-6a95-4261-957f-44b45ea20a02-kube-api-access-l5fnd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.220374 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "90faa05a-6a95-4261-957f-44b45ea20a02" (UID: "90faa05a-6a95-4261-957f-44b45ea20a02"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.234625 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-66d678ddfc-gvbk5"] Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.303792 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.324239 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90faa05a-6a95-4261-957f-44b45ea20a02" (UID: "90faa05a-6a95-4261-957f-44b45ea20a02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.351253 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-config-data" (OuterVolumeSpecName: "config-data") pod "90faa05a-6a95-4261-957f-44b45ea20a02" (UID: "90faa05a-6a95-4261-957f-44b45ea20a02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.405669 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.405703 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90faa05a-6a95-4261-957f-44b45ea20a02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.442119 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.453593 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.470459 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:02:23 crc kubenswrapper[4573]: E1203 09:02:23.470872 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="ceilometer-notification-agent" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.470889 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="ceilometer-notification-agent" Dec 03 09:02:23 crc kubenswrapper[4573]: E1203 09:02:23.470900 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="proxy-httpd" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.470907 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="proxy-httpd" Dec 03 09:02:23 crc kubenswrapper[4573]: E1203 09:02:23.470921 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="sg-core" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.470927 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="sg-core" Dec 03 09:02:23 crc kubenswrapper[4573]: E1203 09:02:23.470953 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="ceilometer-central-agent" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.470959 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="ceilometer-central-agent" Dec 03 09:02:23 crc kubenswrapper[4573]: E1203 09:02:23.470980 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api-log" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.470987 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api-log" Dec 03 09:02:23 crc kubenswrapper[4573]: E1203 09:02:23.470999 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.471005 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.471180 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api-log" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.471197 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="sg-core" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.471206 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="04646290-c4ca-444e-b752-12a054221b16" containerName="cinder-api" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.471217 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="ceilometer-central-agent" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.471231 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="ceilometer-notification-agent" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.471285 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" containerName="proxy-httpd" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.472787 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.480837 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.481159 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.481614 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.500385 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614713 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f75b291-5f0e-422f-a640-a3a231c400ee-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614763 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614810 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614841 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccppl\" (UniqueName: \"kubernetes.io/projected/3f75b291-5f0e-422f-a640-a3a231c400ee-kube-api-access-ccppl\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614861 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614894 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-config-data\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614936 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-scripts\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614956 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f75b291-5f0e-422f-a640-a3a231c400ee-logs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.614999 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.716784 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f75b291-5f0e-422f-a640-a3a231c400ee-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.716840 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.716882 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.716914 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccppl\" (UniqueName: \"kubernetes.io/projected/3f75b291-5f0e-422f-a640-a3a231c400ee-kube-api-access-ccppl\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.716934 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.716967 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-config-data\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.717017 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-scripts\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.717033 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f75b291-5f0e-422f-a640-a3a231c400ee-logs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.717103 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f75b291-5f0e-422f-a640-a3a231c400ee-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.717120 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.718110 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f75b291-5f0e-422f-a640-a3a231c400ee-logs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.724780 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-scripts\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.728397 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.739751 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.740418 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-config-data\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.740585 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.740606 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f75b291-5f0e-422f-a640-a3a231c400ee-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.763041 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccppl\" (UniqueName: \"kubernetes.io/projected/3f75b291-5f0e-422f-a640-a3a231c400ee-kube-api-access-ccppl\") pod \"cinder-api-0\" (UID: \"3f75b291-5f0e-422f-a640-a3a231c400ee\") " pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.796432 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.869123 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"90faa05a-6a95-4261-957f-44b45ea20a02","Type":"ContainerDied","Data":"7d2a7956e549e3a393d94e1c0b347989be53fad2b22a1a0b0f792c9a9704ea69"} Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.869241 4573 scope.go:117] "RemoveContainer" containerID="434e113044392686995debedfeaeba680f9eb5c738e1a1c4235fe23fda564b20" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.869379 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.877460 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"4357c5a1-7325-4bff-8a79-4ffd70879b1c","Type":"ContainerStarted","Data":"b5e6d86a53c7fba744c7b28ff0087a451b1bd7a74c89d7e23593da9d9a446985"} Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.884919 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66d678ddfc-gvbk5" event={"ID":"02bc648a-d9c8-4fdf-bf48-bc57a6b05386","Type":"ContainerStarted","Data":"20b472bb188a0c2da91440418a92baad600120bed405c1a7df313152cd76cc31"} Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.884962 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66d678ddfc-gvbk5" event={"ID":"02bc648a-d9c8-4fdf-bf48-bc57a6b05386","Type":"ContainerStarted","Data":"d06e5f537064eea5a8bddc8890881c1b79d27146c92e76ec884707351a064969"} Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.884971 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66d678ddfc-gvbk5" event={"ID":"02bc648a-d9c8-4fdf-bf48-bc57a6b05386","Type":"ContainerStarted","Data":"c41adad26d9452143bcc29414d4c223ef2e8d71b01f408ec19ef503f57fce349"} Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.906873 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.915594413 podStartE2EDuration="26.906849756s" podCreationTimestamp="2025-12-03 09:01:57 +0000 UTC" firstStartedPulling="2025-12-03 09:01:59.412537852 +0000 UTC m=+1439.980917111" lastFinishedPulling="2025-12-03 09:02:22.403793185 +0000 UTC m=+1462.972172454" observedRunningTime="2025-12-03 09:02:23.897360115 +0000 UTC m=+1464.465739394" watchObservedRunningTime="2025-12-03 09:02:23.906849756 +0000 UTC m=+1464.475229015" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.907406 4573 scope.go:117] "RemoveContainer" containerID="cac51223d774810a3e39697796598123271ab044463433d6e9d852054fe47f66" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.926473 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-66d678ddfc-gvbk5" podStartSLOduration=10.926454256 podStartE2EDuration="10.926454256s" podCreationTimestamp="2025-12-03 09:02:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:02:23.919216947 +0000 UTC m=+1464.487596226" watchObservedRunningTime="2025-12-03 09:02:23.926454256 +0000 UTC m=+1464.494833515" Dec 03 09:02:23 crc kubenswrapper[4573]: I1203 09:02:23.941260 4573 scope.go:117] "RemoveContainer" containerID="0297b7b023135e13bf67973ef377fc64ac67fc725226af9ac25fc6617e699a01" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.025732 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.053900 4573 scope.go:117] "RemoveContainer" containerID="06a981662731fec0c6a3b3b1df80b28b0c540b6de88e6c734f30a5524cee419a" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.248069 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04646290-c4ca-444e-b752-12a054221b16" path="/var/lib/kubelet/pods/04646290-c4ca-444e-b752-12a054221b16/volumes" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.249640 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.259512 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.265745 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.269585 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.269880 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.273162 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.341078 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlpw5\" (UniqueName: \"kubernetes.io/projected/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-kube-api-access-qlpw5\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.341159 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-scripts\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.341220 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.341273 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-config-data\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.341324 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-log-httpd\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.341358 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-run-httpd\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.341388 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.444073 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-log-httpd\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.443521 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-log-httpd\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.444550 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-run-httpd\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.444974 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.444912 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-run-httpd\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.446173 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlpw5\" (UniqueName: \"kubernetes.io/projected/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-kube-api-access-qlpw5\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.446322 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-scripts\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.446483 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.446595 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-config-data\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.453551 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.454186 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-scripts\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.457905 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-config-data\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.462464 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.471592 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlpw5\" (UniqueName: \"kubernetes.io/projected/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-kube-api-access-qlpw5\") pod \"ceilometer-0\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.479474 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.610164 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.972867 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.973831 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerName="glance-log" containerID="cri-o://1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287" gracePeriod=30 Dec 03 09:02:24 crc kubenswrapper[4573]: I1203 09:02:24.974146 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerName="glance-httpd" containerID="cri-o://5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1" gracePeriod=30 Dec 03 09:02:25 crc kubenswrapper[4573]: I1203 09:02:25.032205 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f75b291-5f0e-422f-a640-a3a231c400ee","Type":"ContainerStarted","Data":"4ffdb90919708b1b78730a67638c8edc91dcb2b4f00ee67a1c0c6bced3795202"} Dec 03 09:02:25 crc kubenswrapper[4573]: I1203 09:02:25.032719 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:25 crc kubenswrapper[4573]: I1203 09:02:25.035626 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:25 crc kubenswrapper[4573]: I1203 09:02:25.302812 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:25 crc kubenswrapper[4573]: I1203 09:02:25.340816 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.049781 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90faa05a-6a95-4261-957f-44b45ea20a02" path="/var/lib/kubelet/pods/90faa05a-6a95-4261-957f-44b45ea20a02/volumes" Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.051788 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f75b291-5f0e-422f-a640-a3a231c400ee","Type":"ContainerStarted","Data":"680c919b26a41959294ca202683ced057c5f2b5cbd6bd71b7027b06a5a0c5a62"} Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.051837 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerStarted","Data":"080c22aa258fb40ff8ef58cf0f1ba9108be4ef18fbc198e0f2dd93dfc4d67086"} Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.061453 4573 generic.go:334] "Generic (PLEG): container finished" podID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerID="1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287" exitCode=143 Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.061937 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8f33c171-b5c5-444e-9cd0-63e2a2322f1c","Type":"ContainerDied","Data":"1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287"} Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.943371 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.943915 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.943965 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.944836 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b3b734897ed0efe44e3ec39a1d49da70613f4f9958b9ef5aee9f0572f629605"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:02:26 crc kubenswrapper[4573]: I1203 09:02:26.944902 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://2b3b734897ed0efe44e3ec39a1d49da70613f4f9958b9ef5aee9f0572f629605" gracePeriod=600 Dec 03 09:02:27 crc kubenswrapper[4573]: I1203 09:02:27.073707 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f75b291-5f0e-422f-a640-a3a231c400ee","Type":"ContainerStarted","Data":"4569d91e1f2322b48fe6be83059ad248b9c97a930dfdf7d43e663bb6292a0884"} Dec 03 09:02:27 crc kubenswrapper[4573]: I1203 09:02:27.074146 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 09:02:27 crc kubenswrapper[4573]: I1203 09:02:27.076333 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerStarted","Data":"a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30"} Dec 03 09:02:27 crc kubenswrapper[4573]: I1203 09:02:27.110945 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.110923298 podStartE2EDuration="4.110923298s" podCreationTimestamp="2025-12-03 09:02:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:02:27.103830433 +0000 UTC m=+1467.672209702" watchObservedRunningTime="2025-12-03 09:02:27.110923298 +0000 UTC m=+1467.679302557" Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.088629 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="2b3b734897ed0efe44e3ec39a1d49da70613f4f9958b9ef5aee9f0572f629605" exitCode=0 Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.088693 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"2b3b734897ed0efe44e3ec39a1d49da70613f4f9958b9ef5aee9f0572f629605"} Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.089276 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85"} Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.089306 4573 scope.go:117] "RemoveContainer" containerID="93ffc1f0cd99ae07290bc9e27b4a1d6c063abbb15b748bb21e1aa18b0c38a447" Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.616499 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.617159 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-log" containerID="cri-o://0d7f332fa30855ffec74354857b86688208d19f2199b44d2ebb3d88b00a73383" gracePeriod=30 Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.617311 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-httpd" containerID="cri-o://7f167bafe9560fae3cbca25f2f4f46eefa44a83cac71cba23ba0904e8c68bf7e" gracePeriod=30 Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.763170 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:28 crc kubenswrapper[4573]: I1203 09:02:28.810902 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-66d678ddfc-gvbk5" Dec 03 09:02:29 crc kubenswrapper[4573]: I1203 09:02:29.117310 4573 generic.go:334] "Generic (PLEG): container finished" podID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerID="0d7f332fa30855ffec74354857b86688208d19f2199b44d2ebb3d88b00a73383" exitCode=143 Dec 03 09:02:29 crc kubenswrapper[4573]: I1203 09:02:29.117671 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50863c47-12be-414c-9bd9-fdb5347fea7a","Type":"ContainerDied","Data":"0d7f332fa30855ffec74354857b86688208d19f2199b44d2ebb3d88b00a73383"} Dec 03 09:02:29 crc kubenswrapper[4573]: I1203 09:02:29.127621 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerStarted","Data":"13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd"} Dec 03 09:02:29 crc kubenswrapper[4573]: I1203 09:02:29.127659 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerStarted","Data":"5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6"} Dec 03 09:02:29 crc kubenswrapper[4573]: E1203 09:02:29.304341 4573 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f33c171_b5c5_444e_9cd0_63e2a2322f1c.slice/crio-5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1.scope\": RecentStats: unable to find data in memory cache]" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.120103 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.215608 4573 generic.go:334] "Generic (PLEG): container finished" podID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerID="5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1" exitCode=0 Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.215867 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8f33c171-b5c5-444e-9cd0-63e2a2322f1c","Type":"ContainerDied","Data":"5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1"} Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.215899 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8f33c171-b5c5-444e-9cd0-63e2a2322f1c","Type":"ContainerDied","Data":"196a3fd162473cb4c1ddfb831ba268afae2b1bb7f0c12d97a1fbd7af61aae0a1"} Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.215917 4573 scope.go:117] "RemoveContainer" containerID="5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.216083 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.244747 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkv6s\" (UniqueName: \"kubernetes.io/projected/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-kube-api-access-mkv6s\") pod \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.244813 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-scripts\") pod \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.244930 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-combined-ca-bundle\") pod \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.244980 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-logs\") pod \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.245027 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-config-data\") pod \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.245122 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.245183 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-public-tls-certs\") pod \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.245242 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-httpd-run\") pod \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\" (UID: \"8f33c171-b5c5-444e-9cd0-63e2a2322f1c\") " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.255237 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-kube-api-access-mkv6s" (OuterVolumeSpecName: "kube-api-access-mkv6s") pod "8f33c171-b5c5-444e-9cd0-63e2a2322f1c" (UID: "8f33c171-b5c5-444e-9cd0-63e2a2322f1c"). InnerVolumeSpecName "kube-api-access-mkv6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.255610 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8f33c171-b5c5-444e-9cd0-63e2a2322f1c" (UID: "8f33c171-b5c5-444e-9cd0-63e2a2322f1c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.258691 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "8f33c171-b5c5-444e-9cd0-63e2a2322f1c" (UID: "8f33c171-b5c5-444e-9cd0-63e2a2322f1c"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.258978 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-logs" (OuterVolumeSpecName: "logs") pod "8f33c171-b5c5-444e-9cd0-63e2a2322f1c" (UID: "8f33c171-b5c5-444e-9cd0-63e2a2322f1c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.261687 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-scripts" (OuterVolumeSpecName: "scripts") pod "8f33c171-b5c5-444e-9cd0-63e2a2322f1c" (UID: "8f33c171-b5c5-444e-9cd0-63e2a2322f1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.317923 4573 scope.go:117] "RemoveContainer" containerID="1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.362590 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.362653 4573 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.362672 4573 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.362681 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkv6s\" (UniqueName: \"kubernetes.io/projected/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-kube-api-access-mkv6s\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.362692 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.395981 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f33c171-b5c5-444e-9cd0-63e2a2322f1c" (UID: "8f33c171-b5c5-444e-9cd0-63e2a2322f1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.429510 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8f33c171-b5c5-444e-9cd0-63e2a2322f1c" (UID: "8f33c171-b5c5-444e-9cd0-63e2a2322f1c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.432471 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-config-data" (OuterVolumeSpecName: "config-data") pod "8f33c171-b5c5-444e-9cd0-63e2a2322f1c" (UID: "8f33c171-b5c5-444e-9cd0-63e2a2322f1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.459252 4573 scope.go:117] "RemoveContainer" containerID="5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.463062 4573 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 03 09:02:30 crc kubenswrapper[4573]: E1203 09:02:30.463192 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1\": container with ID starting with 5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1 not found: ID does not exist" containerID="5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.463257 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1"} err="failed to get container status \"5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1\": rpc error: code = NotFound desc = could not find container \"5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1\": container with ID starting with 5339f6be0dbfe29edac8a33778c0db725753dc9ebbe2bac3f546699dc16d3dd1 not found: ID does not exist" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.463293 4573 scope.go:117] "RemoveContainer" containerID="1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.464410 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.464435 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.464445 4573 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.464454 4573 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f33c171-b5c5-444e-9cd0-63e2a2322f1c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:30 crc kubenswrapper[4573]: E1203 09:02:30.466202 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287\": container with ID starting with 1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287 not found: ID does not exist" containerID="1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.466240 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287"} err="failed to get container status \"1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287\": rpc error: code = NotFound desc = could not find container \"1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287\": container with ID starting with 1d182f1b43e5d7da10cb24c1c42f0852c84f495080cd0044d9c3511f5579d287 not found: ID does not exist" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.567115 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.586910 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.613958 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:02:30 crc kubenswrapper[4573]: E1203 09:02:30.614554 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerName="glance-log" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.614580 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerName="glance-log" Dec 03 09:02:30 crc kubenswrapper[4573]: E1203 09:02:30.614605 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerName="glance-httpd" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.614614 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerName="glance-httpd" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.614856 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerName="glance-httpd" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.614895 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" containerName="glance-log" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.616158 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.621031 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.621339 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.628288 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.771754 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-config-data\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.771901 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36bf5497-62bd-4e75-99e8-0a0272cd644a-logs\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.771949 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.772125 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc5q9\" (UniqueName: \"kubernetes.io/projected/36bf5497-62bd-4e75-99e8-0a0272cd644a-kube-api-access-gc5q9\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.772226 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-scripts\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.772268 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36bf5497-62bd-4e75-99e8-0a0272cd644a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.772357 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.772398 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.873850 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-scripts\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.873901 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36bf5497-62bd-4e75-99e8-0a0272cd644a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.873965 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.873998 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.874077 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-config-data\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.874152 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36bf5497-62bd-4e75-99e8-0a0272cd644a-logs\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.874184 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.874232 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc5q9\" (UniqueName: \"kubernetes.io/projected/36bf5497-62bd-4e75-99e8-0a0272cd644a-kube-api-access-gc5q9\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.874339 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.874767 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/36bf5497-62bd-4e75-99e8-0a0272cd644a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.875027 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36bf5497-62bd-4e75-99e8-0a0272cd644a-logs\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.882842 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.891948 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-config-data\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.903107 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-scripts\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.908786 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36bf5497-62bd-4e75-99e8-0a0272cd644a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.911994 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc5q9\" (UniqueName: \"kubernetes.io/projected/36bf5497-62bd-4e75-99e8-0a0272cd644a-kube-api-access-gc5q9\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.921880 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"36bf5497-62bd-4e75-99e8-0a0272cd644a\") " pod="openstack/glance-default-external-api-0" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.924452 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 09:02:30 crc kubenswrapper[4573]: I1203 09:02:30.951272 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 09:02:31 crc kubenswrapper[4573]: I1203 09:02:31.075224 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 03 09:02:31 crc kubenswrapper[4573]: I1203 09:02:31.238586 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerStarted","Data":"e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9"} Dec 03 09:02:31 crc kubenswrapper[4573]: I1203 09:02:31.238956 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:02:31 crc kubenswrapper[4573]: I1203 09:02:31.273961 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.109821893 podStartE2EDuration="8.273944285s" podCreationTimestamp="2025-12-03 09:02:23 +0000 UTC" firstStartedPulling="2025-12-03 09:02:25.340556879 +0000 UTC m=+1465.908936138" lastFinishedPulling="2025-12-03 09:02:30.504679271 +0000 UTC m=+1471.073058530" observedRunningTime="2025-12-03 09:02:31.270577753 +0000 UTC m=+1471.838957012" watchObservedRunningTime="2025-12-03 09:02:31.273944285 +0000 UTC m=+1471.842323544" Dec 03 09:02:31 crc kubenswrapper[4573]: I1203 09:02:31.658630 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.054227 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f33c171-b5c5-444e-9cd0-63e2a2322f1c" path="/var/lib/kubelet/pods/8f33c171-b5c5-444e-9cd0-63e2a2322f1c/volumes" Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.103297 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:52622->10.217.0.153:9292: read: connection reset by peer" Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.103944 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.153:9292/healthcheck\": read tcp 10.217.0.2:52610->10.217.0.153:9292: read: connection reset by peer" Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.295725 4573 generic.go:334] "Generic (PLEG): container finished" podID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerID="7f167bafe9560fae3cbca25f2f4f46eefa44a83cac71cba23ba0904e8c68bf7e" exitCode=0 Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.295879 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50863c47-12be-414c-9bd9-fdb5347fea7a","Type":"ContainerDied","Data":"7f167bafe9560fae3cbca25f2f4f46eefa44a83cac71cba23ba0904e8c68bf7e"} Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.298580 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"36bf5497-62bd-4e75-99e8-0a0272cd644a","Type":"ContainerStarted","Data":"7f229bb67d2679704764cf76cdaa53fffea111102c2b2eba69d4b46e31465c40"} Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.926495 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.930900 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-internal-tls-certs\") pod \"50863c47-12be-414c-9bd9-fdb5347fea7a\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.930981 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-httpd-run\") pod \"50863c47-12be-414c-9bd9-fdb5347fea7a\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.931024 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-config-data\") pod \"50863c47-12be-414c-9bd9-fdb5347fea7a\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.931082 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"50863c47-12be-414c-9bd9-fdb5347fea7a\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.931099 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-scripts\") pod \"50863c47-12be-414c-9bd9-fdb5347fea7a\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.931120 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-logs\") pod \"50863c47-12be-414c-9bd9-fdb5347fea7a\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.931173 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm2dl\" (UniqueName: \"kubernetes.io/projected/50863c47-12be-414c-9bd9-fdb5347fea7a-kube-api-access-sm2dl\") pod \"50863c47-12be-414c-9bd9-fdb5347fea7a\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.931189 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-combined-ca-bundle\") pod \"50863c47-12be-414c-9bd9-fdb5347fea7a\" (UID: \"50863c47-12be-414c-9bd9-fdb5347fea7a\") " Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.933890 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "50863c47-12be-414c-9bd9-fdb5347fea7a" (UID: "50863c47-12be-414c-9bd9-fdb5347fea7a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.934593 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-logs" (OuterVolumeSpecName: "logs") pod "50863c47-12be-414c-9bd9-fdb5347fea7a" (UID: "50863c47-12be-414c-9bd9-fdb5347fea7a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.941508 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "50863c47-12be-414c-9bd9-fdb5347fea7a" (UID: "50863c47-12be-414c-9bd9-fdb5347fea7a"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 09:02:32 crc kubenswrapper[4573]: I1203 09:02:32.954675 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-scripts" (OuterVolumeSpecName: "scripts") pod "50863c47-12be-414c-9bd9-fdb5347fea7a" (UID: "50863c47-12be-414c-9bd9-fdb5347fea7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:32.993288 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50863c47-12be-414c-9bd9-fdb5347fea7a-kube-api-access-sm2dl" (OuterVolumeSpecName: "kube-api-access-sm2dl") pod "50863c47-12be-414c-9bd9-fdb5347fea7a" (UID: "50863c47-12be-414c-9bd9-fdb5347fea7a"). InnerVolumeSpecName "kube-api-access-sm2dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.023322 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.038418 4573 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.038451 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.038462 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.038471 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm2dl\" (UniqueName: \"kubernetes.io/projected/50863c47-12be-414c-9bd9-fdb5347fea7a-kube-api-access-sm2dl\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.038481 4573 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/50863c47-12be-414c-9bd9-fdb5347fea7a-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.088985 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50863c47-12be-414c-9bd9-fdb5347fea7a" (UID: "50863c47-12be-414c-9bd9-fdb5347fea7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.140036 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.147851 4573 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.149380 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "50863c47-12be-414c-9bd9-fdb5347fea7a" (UID: "50863c47-12be-414c-9bd9-fdb5347fea7a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.169361 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-config-data" (OuterVolumeSpecName: "config-data") pod "50863c47-12be-414c-9bd9-fdb5347fea7a" (UID: "50863c47-12be-414c-9bd9-fdb5347fea7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.242031 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.243265 4573 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.243390 4573 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/50863c47-12be-414c-9bd9-fdb5347fea7a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.323308 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"50863c47-12be-414c-9bd9-fdb5347fea7a","Type":"ContainerDied","Data":"59610d98baec5796929f6e447112192b2af6c68199d48d2762725f6ac0d0773e"} Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.324379 4573 scope.go:117] "RemoveContainer" containerID="7f167bafe9560fae3cbca25f2f4f46eefa44a83cac71cba23ba0904e8c68bf7e" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.324597 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.334264 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="ceilometer-central-agent" containerID="cri-o://a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30" gracePeriod=30 Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.334468 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"36bf5497-62bd-4e75-99e8-0a0272cd644a","Type":"ContainerStarted","Data":"834c4f68aafdfdf32f614668bb0995d63af8d35493cdc4f2fdbbacd76640e879"} Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.334982 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="ceilometer-notification-agent" containerID="cri-o://5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6" gracePeriod=30 Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.334542 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="proxy-httpd" containerID="cri-o://e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9" gracePeriod=30 Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.335307 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="sg-core" containerID="cri-o://13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd" gracePeriod=30 Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.409588 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.412732 4573 scope.go:117] "RemoveContainer" containerID="0d7f332fa30855ffec74354857b86688208d19f2199b44d2ebb3d88b00a73383" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.425241 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.444458 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:02:33 crc kubenswrapper[4573]: E1203 09:02:33.445190 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-log" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.445207 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-log" Dec 03 09:02:33 crc kubenswrapper[4573]: E1203 09:02:33.445226 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-httpd" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.445232 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-httpd" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.445443 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-log" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.445461 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" containerName="glance-httpd" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.446519 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.454171 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.458914 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.473968 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.555256 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.555394 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.555432 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.555563 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.555596 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.555967 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.556091 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.556135 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pgx2\" (UniqueName: \"kubernetes.io/projected/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-kube-api-access-7pgx2\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.657651 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.657949 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.658106 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.658194 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.658313 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.658415 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.658512 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pgx2\" (UniqueName: \"kubernetes.io/projected/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-kube-api-access-7pgx2\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.658600 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.663018 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.663198 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.663508 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.677234 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.678864 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.682733 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.686745 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.698002 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.702933 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pgx2\" (UniqueName: \"kubernetes.io/projected/10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2-kube-api-access-7pgx2\") pod \"glance-default-internal-api-0\" (UID: \"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2\") " pod="openstack/glance-default-internal-api-0" Dec 03 09:02:33 crc kubenswrapper[4573]: I1203 09:02:33.792523 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:34 crc kubenswrapper[4573]: I1203 09:02:34.076566 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50863c47-12be-414c-9bd9-fdb5347fea7a" path="/var/lib/kubelet/pods/50863c47-12be-414c-9bd9-fdb5347fea7a/volumes" Dec 03 09:02:34 crc kubenswrapper[4573]: I1203 09:02:34.354848 4573 generic.go:334] "Generic (PLEG): container finished" podID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerID="e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9" exitCode=0 Dec 03 09:02:34 crc kubenswrapper[4573]: I1203 09:02:34.356498 4573 generic.go:334] "Generic (PLEG): container finished" podID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerID="13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd" exitCode=2 Dec 03 09:02:34 crc kubenswrapper[4573]: I1203 09:02:34.355160 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerDied","Data":"e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9"} Dec 03 09:02:34 crc kubenswrapper[4573]: I1203 09:02:34.359415 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerDied","Data":"13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd"} Dec 03 09:02:34 crc kubenswrapper[4573]: I1203 09:02:34.378896 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"36bf5497-62bd-4e75-99e8-0a0272cd644a","Type":"ContainerStarted","Data":"30fb5f8c62cb03c53639fe30610fd104afba615b700ec1802e0ae9ed80fe2972"} Dec 03 09:02:34 crc kubenswrapper[4573]: I1203 09:02:34.425159 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.425139992 podStartE2EDuration="4.425139992s" podCreationTimestamp="2025-12-03 09:02:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:02:34.404294598 +0000 UTC m=+1474.972673857" watchObservedRunningTime="2025-12-03 09:02:34.425139992 +0000 UTC m=+1474.993519251" Dec 03 09:02:34 crc kubenswrapper[4573]: I1203 09:02:34.611512 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 09:02:35 crc kubenswrapper[4573]: I1203 09:02:35.390456 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2","Type":"ContainerStarted","Data":"3cb2d45ddf0a4235e61a1655b4a5e140d8338a013cfe72a7e75489c656e00ad7"} Dec 03 09:02:35 crc kubenswrapper[4573]: I1203 09:02:35.396081 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerDied","Data":"5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6"} Dec 03 09:02:35 crc kubenswrapper[4573]: I1203 09:02:35.396040 4573 generic.go:334] "Generic (PLEG): container finished" podID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerID="5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6" exitCode=0 Dec 03 09:02:36 crc kubenswrapper[4573]: I1203 09:02:36.409321 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2","Type":"ContainerStarted","Data":"9960dd9319483054a387e5503adcfd2d25db4938f636a59825ff88045cc01b4e"} Dec 03 09:02:36 crc kubenswrapper[4573]: I1203 09:02:36.409577 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2","Type":"ContainerStarted","Data":"93cdc066f0315325c39e8f356db2ef26a43d14a7faa6c732b09206835d764080"} Dec 03 09:02:36 crc kubenswrapper[4573]: I1203 09:02:36.454004 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.453984596 podStartE2EDuration="3.453984596s" podCreationTimestamp="2025-12-03 09:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:02:36.445323057 +0000 UTC m=+1477.013702316" watchObservedRunningTime="2025-12-03 09:02:36.453984596 +0000 UTC m=+1477.022363855" Dec 03 09:02:36 crc kubenswrapper[4573]: I1203 09:02:36.959727 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.033437 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-run-httpd\") pod \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.033506 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-config-data\") pod \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.033550 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-log-httpd\") pod \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.033624 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-combined-ca-bundle\") pod \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.033731 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-sg-core-conf-yaml\") pod \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.033779 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-scripts\") pod \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.033826 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlpw5\" (UniqueName: \"kubernetes.io/projected/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-kube-api-access-qlpw5\") pod \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\" (UID: \"863f70ec-b02e-4e42-8c2e-85f009ec8d5d\") " Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.034531 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "863f70ec-b02e-4e42-8c2e-85f009ec8d5d" (UID: "863f70ec-b02e-4e42-8c2e-85f009ec8d5d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.034574 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "863f70ec-b02e-4e42-8c2e-85f009ec8d5d" (UID: "863f70ec-b02e-4e42-8c2e-85f009ec8d5d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.034758 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.034772 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.046269 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-kube-api-access-qlpw5" (OuterVolumeSpecName: "kube-api-access-qlpw5") pod "863f70ec-b02e-4e42-8c2e-85f009ec8d5d" (UID: "863f70ec-b02e-4e42-8c2e-85f009ec8d5d"). InnerVolumeSpecName "kube-api-access-qlpw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.058233 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-scripts" (OuterVolumeSpecName: "scripts") pod "863f70ec-b02e-4e42-8c2e-85f009ec8d5d" (UID: "863f70ec-b02e-4e42-8c2e-85f009ec8d5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.137098 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.137145 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlpw5\" (UniqueName: \"kubernetes.io/projected/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-kube-api-access-qlpw5\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.198502 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "863f70ec-b02e-4e42-8c2e-85f009ec8d5d" (UID: "863f70ec-b02e-4e42-8c2e-85f009ec8d5d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.240552 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.251611 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "863f70ec-b02e-4e42-8c2e-85f009ec8d5d" (UID: "863f70ec-b02e-4e42-8c2e-85f009ec8d5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.271566 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-config-data" (OuterVolumeSpecName: "config-data") pod "863f70ec-b02e-4e42-8c2e-85f009ec8d5d" (UID: "863f70ec-b02e-4e42-8c2e-85f009ec8d5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.342430 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.342472 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863f70ec-b02e-4e42-8c2e-85f009ec8d5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.420136 4573 generic.go:334] "Generic (PLEG): container finished" podID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerID="a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30" exitCode=0 Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.421135 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.425291 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerDied","Data":"a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30"} Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.425338 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863f70ec-b02e-4e42-8c2e-85f009ec8d5d","Type":"ContainerDied","Data":"080c22aa258fb40ff8ef58cf0f1ba9108be4ef18fbc198e0f2dd93dfc4d67086"} Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.425356 4573 scope.go:117] "RemoveContainer" containerID="e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.482499 4573 scope.go:117] "RemoveContainer" containerID="13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.544284 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.547088 4573 scope.go:117] "RemoveContainer" containerID="5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.555654 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.581252 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:37 crc kubenswrapper[4573]: E1203 09:02:37.581723 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="ceilometer-central-agent" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.581741 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="ceilometer-central-agent" Dec 03 09:02:37 crc kubenswrapper[4573]: E1203 09:02:37.581755 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="ceilometer-notification-agent" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.581762 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="ceilometer-notification-agent" Dec 03 09:02:37 crc kubenswrapper[4573]: E1203 09:02:37.581781 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="proxy-httpd" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.581789 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="proxy-httpd" Dec 03 09:02:37 crc kubenswrapper[4573]: E1203 09:02:37.581803 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="sg-core" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.581810 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="sg-core" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.582014 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="ceilometer-notification-agent" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.582039 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="proxy-httpd" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.582078 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="sg-core" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.582095 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" containerName="ceilometer-central-agent" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.583802 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.585940 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.586220 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.601428 4573 scope.go:117] "RemoveContainer" containerID="a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.607256 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.648467 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-config-data\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.648565 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-scripts\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.648633 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.648653 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-log-httpd\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.648703 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b999h\" (UniqueName: \"kubernetes.io/projected/b35cb6f0-c844-4f2f-aa54-b01e439534a1-kube-api-access-b999h\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.648736 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.648785 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-run-httpd\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.664837 4573 scope.go:117] "RemoveContainer" containerID="e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9" Dec 03 09:02:37 crc kubenswrapper[4573]: E1203 09:02:37.665640 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9\": container with ID starting with e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9 not found: ID does not exist" containerID="e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.665736 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9"} err="failed to get container status \"e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9\": rpc error: code = NotFound desc = could not find container \"e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9\": container with ID starting with e5a2a24296878dbe953616a9dba9531bcfa7407b851d6b1323d159c8be3377b9 not found: ID does not exist" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.665803 4573 scope.go:117] "RemoveContainer" containerID="13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd" Dec 03 09:02:37 crc kubenswrapper[4573]: E1203 09:02:37.667951 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd\": container with ID starting with 13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd not found: ID does not exist" containerID="13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.668151 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd"} err="failed to get container status \"13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd\": rpc error: code = NotFound desc = could not find container \"13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd\": container with ID starting with 13f687d729929b461877e7cac06ba6d2a8a0f93c9797b0569ea62559d4bf39fd not found: ID does not exist" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.668250 4573 scope.go:117] "RemoveContainer" containerID="5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6" Dec 03 09:02:37 crc kubenswrapper[4573]: E1203 09:02:37.670446 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6\": container with ID starting with 5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6 not found: ID does not exist" containerID="5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.670518 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6"} err="failed to get container status \"5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6\": rpc error: code = NotFound desc = could not find container \"5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6\": container with ID starting with 5c126de74fee3ccca7dbda181d3296344f71e705f70320b017947ee2943e23b6 not found: ID does not exist" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.670582 4573 scope.go:117] "RemoveContainer" containerID="a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30" Dec 03 09:02:37 crc kubenswrapper[4573]: E1203 09:02:37.672823 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30\": container with ID starting with a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30 not found: ID does not exist" containerID="a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.672942 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30"} err="failed to get container status \"a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30\": rpc error: code = NotFound desc = could not find container \"a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30\": container with ID starting with a606c949f5f7109a704636d38f99f951fca5d81e57913494fa4c727ab2fa9e30 not found: ID does not exist" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.709687 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.759789 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.759853 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-log-httpd\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.759971 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b999h\" (UniqueName: \"kubernetes.io/projected/b35cb6f0-c844-4f2f-aa54-b01e439534a1-kube-api-access-b999h\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.760019 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.760177 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-run-httpd\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.760348 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-config-data\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.760483 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-scripts\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.766212 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-run-httpd\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.775471 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.775760 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-log-httpd\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.786641 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-config-data\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.805398 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.813713 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-scripts\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.816442 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="3f75b291-5f0e-422f-a640-a3a231c400ee" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.822290 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b999h\" (UniqueName: \"kubernetes.io/projected/b35cb6f0-c844-4f2f-aa54-b01e439534a1-kube-api-access-b999h\") pod \"ceilometer-0\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " pod="openstack/ceilometer-0" Dec 03 09:02:37 crc kubenswrapper[4573]: I1203 09:02:37.912959 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:38 crc kubenswrapper[4573]: I1203 09:02:38.071827 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="863f70ec-b02e-4e42-8c2e-85f009ec8d5d" path="/var/lib/kubelet/pods/863f70ec-b02e-4e42-8c2e-85f009ec8d5d/volumes" Dec 03 09:02:38 crc kubenswrapper[4573]: I1203 09:02:38.529132 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:39 crc kubenswrapper[4573]: I1203 09:02:39.443868 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerStarted","Data":"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39"} Dec 03 09:02:39 crc kubenswrapper[4573]: I1203 09:02:39.444215 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerStarted","Data":"9e532848de986c4538420ec3ea85f2f4e00895e6918f3494257baa1b8fc48163"} Dec 03 09:02:40 crc kubenswrapper[4573]: I1203 09:02:40.453957 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerStarted","Data":"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80"} Dec 03 09:02:40 crc kubenswrapper[4573]: I1203 09:02:40.919110 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 09:02:40 crc kubenswrapper[4573]: I1203 09:02:40.919428 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:02:40 crc kubenswrapper[4573]: I1203 09:02:40.920294 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"b8aa7ba7c95bd0a8692253b9b93fecbaa60bed28d93ea2c5355468e1a8ef3047"} pod="openstack/horizon-7b8b5b6556-4zxxs" containerMessage="Container horizon failed startup probe, will be restarted" Dec 03 09:02:40 crc kubenswrapper[4573]: I1203 09:02:40.920336 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" containerID="cri-o://b8aa7ba7c95bd0a8692253b9b93fecbaa60bed28d93ea2c5355468e1a8ef3047" gracePeriod=30 Dec 03 09:02:40 crc kubenswrapper[4573]: I1203 09:02:40.951858 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 09:02:40 crc kubenswrapper[4573]: I1203 09:02:40.951924 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.006992 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.020240 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.071679 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.071755 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.072684 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"eee5245aaff562e364c268a895dd3a53869d1cf3cec162d72a6fbfa629898941"} pod="openstack/horizon-66f99c579b-gh7sh" containerMessage="Container horizon failed startup probe, will be restarted" Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.072722 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" containerID="cri-o://eee5245aaff562e364c268a895dd3a53869d1cf3cec162d72a6fbfa629898941" gracePeriod=30 Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.464822 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerStarted","Data":"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9"} Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.464915 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 09:02:41 crc kubenswrapper[4573]: I1203 09:02:41.465099 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 09:02:42 crc kubenswrapper[4573]: I1203 09:02:42.236635 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.486835 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerStarted","Data":"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9"} Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.487262 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="ceilometer-central-agent" containerID="cri-o://507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39" gracePeriod=30 Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.487492 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.487546 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="proxy-httpd" containerID="cri-o://242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9" gracePeriod=30 Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.487597 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="sg-core" containerID="cri-o://b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9" gracePeriod=30 Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.487670 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="ceilometer-notification-agent" containerID="cri-o://11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80" gracePeriod=30 Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.511097 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5971853019999998 podStartE2EDuration="6.511076052s" podCreationTimestamp="2025-12-03 09:02:37 +0000 UTC" firstStartedPulling="2025-12-03 09:02:38.535481929 +0000 UTC m=+1479.103861188" lastFinishedPulling="2025-12-03 09:02:42.449372679 +0000 UTC m=+1483.017751938" observedRunningTime="2025-12-03 09:02:43.508966034 +0000 UTC m=+1484.077345313" watchObservedRunningTime="2025-12-03 09:02:43.511076052 +0000 UTC m=+1484.079455311" Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.793289 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.793340 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.838012 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:43 crc kubenswrapper[4573]: I1203 09:02:43.874600 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.393414 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.434000 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-combined-ca-bundle\") pod \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.434091 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b999h\" (UniqueName: \"kubernetes.io/projected/b35cb6f0-c844-4f2f-aa54-b01e439534a1-kube-api-access-b999h\") pod \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.434240 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-run-httpd\") pod \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.434300 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-scripts\") pod \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.434320 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-config-data\") pod \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.434337 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-log-httpd\") pod \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.434369 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-sg-core-conf-yaml\") pod \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\" (UID: \"b35cb6f0-c844-4f2f-aa54-b01e439534a1\") " Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.435453 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b35cb6f0-c844-4f2f-aa54-b01e439534a1" (UID: "b35cb6f0-c844-4f2f-aa54-b01e439534a1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.435636 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b35cb6f0-c844-4f2f-aa54-b01e439534a1" (UID: "b35cb6f0-c844-4f2f-aa54-b01e439534a1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.444142 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b35cb6f0-c844-4f2f-aa54-b01e439534a1-kube-api-access-b999h" (OuterVolumeSpecName: "kube-api-access-b999h") pod "b35cb6f0-c844-4f2f-aa54-b01e439534a1" (UID: "b35cb6f0-c844-4f2f-aa54-b01e439534a1"). InnerVolumeSpecName "kube-api-access-b999h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.446564 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-scripts" (OuterVolumeSpecName: "scripts") pod "b35cb6f0-c844-4f2f-aa54-b01e439534a1" (UID: "b35cb6f0-c844-4f2f-aa54-b01e439534a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.481949 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b35cb6f0-c844-4f2f-aa54-b01e439534a1" (UID: "b35cb6f0-c844-4f2f-aa54-b01e439534a1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.504263 4573 generic.go:334] "Generic (PLEG): container finished" podID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerID="242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9" exitCode=0 Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.505549 4573 generic.go:334] "Generic (PLEG): container finished" podID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerID="b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9" exitCode=2 Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.505806 4573 generic.go:334] "Generic (PLEG): container finished" podID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerID="11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80" exitCode=0 Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.505937 4573 generic.go:334] "Generic (PLEG): container finished" podID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerID="507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39" exitCode=0 Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.507610 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerDied","Data":"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9"} Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.507817 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerDied","Data":"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9"} Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.507934 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.508061 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.508763 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerDied","Data":"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80"} Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.508886 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerDied","Data":"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39"} Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.509425 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35cb6f0-c844-4f2f-aa54-b01e439534a1","Type":"ContainerDied","Data":"9e532848de986c4538420ec3ea85f2f4e00895e6918f3494257baa1b8fc48163"} Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.508736 4573 scope.go:117] "RemoveContainer" containerID="242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.508303 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.536440 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b999h\" (UniqueName: \"kubernetes.io/projected/b35cb6f0-c844-4f2f-aa54-b01e439534a1-kube-api-access-b999h\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.536470 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.536481 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.536490 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35cb6f0-c844-4f2f-aa54-b01e439534a1-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.536501 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.547767 4573 scope.go:117] "RemoveContainer" containerID="b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.566024 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b35cb6f0-c844-4f2f-aa54-b01e439534a1" (UID: "b35cb6f0-c844-4f2f-aa54-b01e439534a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.573232 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-config-data" (OuterVolumeSpecName: "config-data") pod "b35cb6f0-c844-4f2f-aa54-b01e439534a1" (UID: "b35cb6f0-c844-4f2f-aa54-b01e439534a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.574424 4573 scope.go:117] "RemoveContainer" containerID="11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.604397 4573 scope.go:117] "RemoveContainer" containerID="507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.639165 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.639209 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35cb6f0-c844-4f2f-aa54-b01e439534a1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.708326 4573 scope.go:117] "RemoveContainer" containerID="242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9" Dec 03 09:02:44 crc kubenswrapper[4573]: E1203 09:02:44.712001 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": container with ID starting with 242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9 not found: ID does not exist" containerID="242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.712087 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9"} err="failed to get container status \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": rpc error: code = NotFound desc = could not find container \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": container with ID starting with 242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.712117 4573 scope.go:117] "RemoveContainer" containerID="b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9" Dec 03 09:02:44 crc kubenswrapper[4573]: E1203 09:02:44.712878 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": container with ID starting with b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9 not found: ID does not exist" containerID="b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.712900 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9"} err="failed to get container status \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": rpc error: code = NotFound desc = could not find container \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": container with ID starting with b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.712916 4573 scope.go:117] "RemoveContainer" containerID="11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80" Dec 03 09:02:44 crc kubenswrapper[4573]: E1203 09:02:44.713337 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": container with ID starting with 11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80 not found: ID does not exist" containerID="11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.713359 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80"} err="failed to get container status \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": rpc error: code = NotFound desc = could not find container \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": container with ID starting with 11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.713373 4573 scope.go:117] "RemoveContainer" containerID="507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39" Dec 03 09:02:44 crc kubenswrapper[4573]: E1203 09:02:44.713714 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": container with ID starting with 507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39 not found: ID does not exist" containerID="507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.713739 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39"} err="failed to get container status \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": rpc error: code = NotFound desc = could not find container \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": container with ID starting with 507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.713761 4573 scope.go:117] "RemoveContainer" containerID="242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.714147 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9"} err="failed to get container status \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": rpc error: code = NotFound desc = could not find container \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": container with ID starting with 242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.714169 4573 scope.go:117] "RemoveContainer" containerID="b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.714472 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9"} err="failed to get container status \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": rpc error: code = NotFound desc = could not find container \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": container with ID starting with b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.714490 4573 scope.go:117] "RemoveContainer" containerID="11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.714820 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80"} err="failed to get container status \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": rpc error: code = NotFound desc = could not find container \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": container with ID starting with 11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.714842 4573 scope.go:117] "RemoveContainer" containerID="507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.715177 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39"} err="failed to get container status \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": rpc error: code = NotFound desc = could not find container \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": container with ID starting with 507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.715204 4573 scope.go:117] "RemoveContainer" containerID="242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.715572 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9"} err="failed to get container status \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": rpc error: code = NotFound desc = could not find container \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": container with ID starting with 242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.715598 4573 scope.go:117] "RemoveContainer" containerID="b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.715950 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9"} err="failed to get container status \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": rpc error: code = NotFound desc = could not find container \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": container with ID starting with b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.715974 4573 scope.go:117] "RemoveContainer" containerID="11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.716352 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80"} err="failed to get container status \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": rpc error: code = NotFound desc = could not find container \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": container with ID starting with 11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.716371 4573 scope.go:117] "RemoveContainer" containerID="507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.716664 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39"} err="failed to get container status \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": rpc error: code = NotFound desc = could not find container \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": container with ID starting with 507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.716681 4573 scope.go:117] "RemoveContainer" containerID="242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.717004 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9"} err="failed to get container status \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": rpc error: code = NotFound desc = could not find container \"242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9\": container with ID starting with 242cb4b4ca7b3e10132de1e4a0df60cce9de2c4d71f269a4bd5d907307b2bbe9 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.717027 4573 scope.go:117] "RemoveContainer" containerID="b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.717433 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9"} err="failed to get container status \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": rpc error: code = NotFound desc = could not find container \"b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9\": container with ID starting with b86f67fcf1db85ac68753c61f15f71518e4d6511b06cf776dd8e03bd95d85ca9 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.717449 4573 scope.go:117] "RemoveContainer" containerID="11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.717773 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80"} err="failed to get container status \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": rpc error: code = NotFound desc = could not find container \"11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80\": container with ID starting with 11a30f9e9360939619c8179ce2eae0bd7b85f4897cb75eea4c7d129794395c80 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.717795 4573 scope.go:117] "RemoveContainer" containerID="507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.718149 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39"} err="failed to get container status \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": rpc error: code = NotFound desc = could not find container \"507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39\": container with ID starting with 507c73d2ac1f372fbd075d95b9d016341f0b1ed6c3a6acde77b0165085d5ac39 not found: ID does not exist" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.853257 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.860961 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.881543 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:44 crc kubenswrapper[4573]: E1203 09:02:44.881932 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="ceilometer-notification-agent" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.881960 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="ceilometer-notification-agent" Dec 03 09:02:44 crc kubenswrapper[4573]: E1203 09:02:44.881976 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="proxy-httpd" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.881982 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="proxy-httpd" Dec 03 09:02:44 crc kubenswrapper[4573]: E1203 09:02:44.881994 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="sg-core" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.882002 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="sg-core" Dec 03 09:02:44 crc kubenswrapper[4573]: E1203 09:02:44.882028 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="ceilometer-central-agent" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.882035 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="ceilometer-central-agent" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.882233 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="sg-core" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.882250 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="ceilometer-notification-agent" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.882261 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="proxy-httpd" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.882273 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" containerName="ceilometer-central-agent" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.883862 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.892063 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.892362 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.913711 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.973933 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.974078 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 09:02:44 crc kubenswrapper[4573]: I1203 09:02:44.976124 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.051685 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mcdd\" (UniqueName: \"kubernetes.io/projected/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-kube-api-access-5mcdd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.051738 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-scripts\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.051760 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-log-httpd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.051824 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-run-httpd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.051851 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.051918 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-config-data\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.051959 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.155557 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-run-httpd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.154906 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-run-httpd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.155781 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.157017 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-config-data\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.157192 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.157268 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mcdd\" (UniqueName: \"kubernetes.io/projected/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-kube-api-access-5mcdd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.157330 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-scripts\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.157358 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-log-httpd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.157893 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-log-httpd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.164143 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.168376 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.169314 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-config-data\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.180713 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-scripts\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.183707 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mcdd\" (UniqueName: \"kubernetes.io/projected/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-kube-api-access-5mcdd\") pod \"ceilometer-0\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.213795 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:45 crc kubenswrapper[4573]: I1203 09:02:45.762738 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:46 crc kubenswrapper[4573]: I1203 09:02:46.040469 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b35cb6f0-c844-4f2f-aa54-b01e439534a1" path="/var/lib/kubelet/pods/b35cb6f0-c844-4f2f-aa54-b01e439534a1/volumes" Dec 03 09:02:46 crc kubenswrapper[4573]: I1203 09:02:46.543744 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 09:02:46 crc kubenswrapper[4573]: I1203 09:02:46.544153 4573 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 09:02:46 crc kubenswrapper[4573]: I1203 09:02:46.545275 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerStarted","Data":"7754d679276903ff33d54bd20e99fe960e5cead0a3c87a8362290fd444e546a7"} Dec 03 09:02:47 crc kubenswrapper[4573]: I1203 09:02:47.092941 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:47 crc kubenswrapper[4573]: I1203 09:02:47.103118 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 09:02:47 crc kubenswrapper[4573]: I1203 09:02:47.553944 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerStarted","Data":"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c"} Dec 03 09:02:47 crc kubenswrapper[4573]: I1203 09:02:47.675420 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:48 crc kubenswrapper[4573]: I1203 09:02:48.567266 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerStarted","Data":"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f"} Dec 03 09:02:49 crc kubenswrapper[4573]: I1203 09:02:49.581121 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerStarted","Data":"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246"} Dec 03 09:02:50 crc kubenswrapper[4573]: I1203 09:02:50.593501 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerStarted","Data":"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b"} Dec 03 09:02:50 crc kubenswrapper[4573]: I1203 09:02:50.593685 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="ceilometer-central-agent" containerID="cri-o://2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c" gracePeriod=30 Dec 03 09:02:50 crc kubenswrapper[4573]: I1203 09:02:50.593805 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="proxy-httpd" containerID="cri-o://ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b" gracePeriod=30 Dec 03 09:02:50 crc kubenswrapper[4573]: I1203 09:02:50.593790 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="ceilometer-notification-agent" containerID="cri-o://658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f" gracePeriod=30 Dec 03 09:02:50 crc kubenswrapper[4573]: I1203 09:02:50.593749 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="sg-core" containerID="cri-o://e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246" gracePeriod=30 Dec 03 09:02:50 crc kubenswrapper[4573]: I1203 09:02:50.593837 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:02:50 crc kubenswrapper[4573]: I1203 09:02:50.679161 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.777022378 podStartE2EDuration="6.679090939s" podCreationTimestamp="2025-12-03 09:02:44 +0000 UTC" firstStartedPulling="2025-12-03 09:02:45.778863983 +0000 UTC m=+1486.347243242" lastFinishedPulling="2025-12-03 09:02:49.680932534 +0000 UTC m=+1490.249311803" observedRunningTime="2025-12-03 09:02:50.624434127 +0000 UTC m=+1491.192813386" watchObservedRunningTime="2025-12-03 09:02:50.679090939 +0000 UTC m=+1491.247470208" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.397489 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.493276 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-log-httpd\") pod \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.493319 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-sg-core-conf-yaml\") pod \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.493428 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-combined-ca-bundle\") pod \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.493461 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mcdd\" (UniqueName: \"kubernetes.io/projected/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-kube-api-access-5mcdd\") pod \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.493502 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-config-data\") pod \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.493616 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-scripts\") pod \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.493756 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-run-httpd\") pod \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\" (UID: \"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7\") " Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.493907 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" (UID: "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.494314 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" (UID: "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.494838 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.494860 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.502604 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-scripts" (OuterVolumeSpecName: "scripts") pod "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" (UID: "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.503335 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-kube-api-access-5mcdd" (OuterVolumeSpecName: "kube-api-access-5mcdd") pod "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" (UID: "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7"). InnerVolumeSpecName "kube-api-access-5mcdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.548189 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" (UID: "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.594319 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" (UID: "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.596902 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.596933 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.596943 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.596953 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mcdd\" (UniqueName: \"kubernetes.io/projected/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-kube-api-access-5mcdd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.607935 4573 generic.go:334] "Generic (PLEG): container finished" podID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerID="ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b" exitCode=0 Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.607996 4573 generic.go:334] "Generic (PLEG): container finished" podID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerID="e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246" exitCode=2 Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608006 4573 generic.go:334] "Generic (PLEG): container finished" podID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerID="658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f" exitCode=0 Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608013 4573 generic.go:334] "Generic (PLEG): container finished" podID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerID="2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c" exitCode=0 Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608006 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerDied","Data":"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b"} Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608114 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608231 4573 scope.go:117] "RemoveContainer" containerID="ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608326 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerDied","Data":"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246"} Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608352 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerDied","Data":"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f"} Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608364 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerDied","Data":"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c"} Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.608378 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7","Type":"ContainerDied","Data":"7754d679276903ff33d54bd20e99fe960e5cead0a3c87a8362290fd444e546a7"} Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.624194 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-config-data" (OuterVolumeSpecName: "config-data") pod "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" (UID: "3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.632313 4573 scope.go:117] "RemoveContainer" containerID="e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.654533 4573 scope.go:117] "RemoveContainer" containerID="658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.687943 4573 scope.go:117] "RemoveContainer" containerID="2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.700366 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.714543 4573 scope.go:117] "RemoveContainer" containerID="ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b" Dec 03 09:02:51 crc kubenswrapper[4573]: E1203 09:02:51.715550 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": container with ID starting with ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b not found: ID does not exist" containerID="ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.715618 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b"} err="failed to get container status \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": rpc error: code = NotFound desc = could not find container \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": container with ID starting with ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.715664 4573 scope.go:117] "RemoveContainer" containerID="e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246" Dec 03 09:02:51 crc kubenswrapper[4573]: E1203 09:02:51.716150 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": container with ID starting with e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246 not found: ID does not exist" containerID="e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.716196 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246"} err="failed to get container status \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": rpc error: code = NotFound desc = could not find container \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": container with ID starting with e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246 not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.716227 4573 scope.go:117] "RemoveContainer" containerID="658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f" Dec 03 09:02:51 crc kubenswrapper[4573]: E1203 09:02:51.716533 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": container with ID starting with 658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f not found: ID does not exist" containerID="658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.716561 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f"} err="failed to get container status \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": rpc error: code = NotFound desc = could not find container \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": container with ID starting with 658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.716578 4573 scope.go:117] "RemoveContainer" containerID="2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c" Dec 03 09:02:51 crc kubenswrapper[4573]: E1203 09:02:51.717107 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": container with ID starting with 2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c not found: ID does not exist" containerID="2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.717133 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c"} err="failed to get container status \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": rpc error: code = NotFound desc = could not find container \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": container with ID starting with 2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.717148 4573 scope.go:117] "RemoveContainer" containerID="ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.717356 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b"} err="failed to get container status \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": rpc error: code = NotFound desc = could not find container \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": container with ID starting with ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.717377 4573 scope.go:117] "RemoveContainer" containerID="e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.718271 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246"} err="failed to get container status \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": rpc error: code = NotFound desc = could not find container \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": container with ID starting with e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246 not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.718311 4573 scope.go:117] "RemoveContainer" containerID="658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.718580 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f"} err="failed to get container status \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": rpc error: code = NotFound desc = could not find container \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": container with ID starting with 658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.718624 4573 scope.go:117] "RemoveContainer" containerID="2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.718856 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c"} err="failed to get container status \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": rpc error: code = NotFound desc = could not find container \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": container with ID starting with 2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.718882 4573 scope.go:117] "RemoveContainer" containerID="ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.719204 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b"} err="failed to get container status \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": rpc error: code = NotFound desc = could not find container \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": container with ID starting with ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.719234 4573 scope.go:117] "RemoveContainer" containerID="e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.719642 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246"} err="failed to get container status \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": rpc error: code = NotFound desc = could not find container \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": container with ID starting with e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246 not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.719670 4573 scope.go:117] "RemoveContainer" containerID="658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.720116 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f"} err="failed to get container status \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": rpc error: code = NotFound desc = could not find container \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": container with ID starting with 658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.720152 4573 scope.go:117] "RemoveContainer" containerID="2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.720524 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c"} err="failed to get container status \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": rpc error: code = NotFound desc = could not find container \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": container with ID starting with 2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.720551 4573 scope.go:117] "RemoveContainer" containerID="ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.720779 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b"} err="failed to get container status \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": rpc error: code = NotFound desc = could not find container \"ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b\": container with ID starting with ec10d98bc93e9ba47353cca4835b372438a6a7dfc1e3a0751e6f9ce592cf3f6b not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.720814 4573 scope.go:117] "RemoveContainer" containerID="e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.721220 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246"} err="failed to get container status \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": rpc error: code = NotFound desc = could not find container \"e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246\": container with ID starting with e0d2a566630bb19f1beeb4944a8a9c82187104ad25ea65715bb38710763dc246 not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.721247 4573 scope.go:117] "RemoveContainer" containerID="658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.721775 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f"} err="failed to get container status \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": rpc error: code = NotFound desc = could not find container \"658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f\": container with ID starting with 658f01fd82d62eb219729e21c4cc27ce90bdce659e2e9be04bfda148e40ceb6f not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.721803 4573 scope.go:117] "RemoveContainer" containerID="2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.722000 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c"} err="failed to get container status \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": rpc error: code = NotFound desc = could not find container \"2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c\": container with ID starting with 2f07161edaaf21f1acd4c5dd2ac72ba2e29d0fa5dec0dd0e4c1398d2e6687a6c not found: ID does not exist" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.943266 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.958668 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.989319 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:51 crc kubenswrapper[4573]: E1203 09:02:51.989829 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="proxy-httpd" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.989854 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="proxy-httpd" Dec 03 09:02:51 crc kubenswrapper[4573]: E1203 09:02:51.989880 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="sg-core" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.989891 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="sg-core" Dec 03 09:02:51 crc kubenswrapper[4573]: E1203 09:02:51.989905 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="ceilometer-notification-agent" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.989913 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="ceilometer-notification-agent" Dec 03 09:02:51 crc kubenswrapper[4573]: E1203 09:02:51.989924 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="ceilometer-central-agent" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.989932 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="ceilometer-central-agent" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.990196 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="proxy-httpd" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.990230 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="ceilometer-central-agent" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.990245 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="sg-core" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.990267 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" containerName="ceilometer-notification-agent" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.991991 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:51 crc kubenswrapper[4573]: I1203 09:02:51.996944 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.008955 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-2kcns"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.010337 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.011030 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.028620 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.043006 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7" path="/var/lib/kubelet/pods/3e3a55f0-91fd-43f8-a3ab-504e8c1a7ff7/volumes" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.044130 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2kcns"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115001 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpznz\" (UniqueName: \"kubernetes.io/projected/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-kube-api-access-lpznz\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115064 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-scripts\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115121 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115145 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgqdq\" (UniqueName: \"kubernetes.io/projected/7ff27d27-5fa2-401b-95a2-c633aa025759-kube-api-access-fgqdq\") pod \"nova-api-db-create-2kcns\" (UID: \"7ff27d27-5fa2-401b-95a2-c633aa025759\") " pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115176 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-run-httpd\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115231 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-log-httpd\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115283 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-config-data\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115328 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff27d27-5fa2-401b-95a2-c633aa025759-operator-scripts\") pod \"nova-api-db-create-2kcns\" (UID: \"7ff27d27-5fa2-401b-95a2-c633aa025759\") " pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.115363 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.125724 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-kp7fb"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.130630 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.139234 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-kp7fb"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.217850 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.218331 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgqdq\" (UniqueName: \"kubernetes.io/projected/7ff27d27-5fa2-401b-95a2-c633aa025759-kube-api-access-fgqdq\") pod \"nova-api-db-create-2kcns\" (UID: \"7ff27d27-5fa2-401b-95a2-c633aa025759\") " pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.218448 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-run-httpd\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.218568 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-log-httpd\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.218682 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/539b8651-7e63-416c-8428-83cd2e4b3813-operator-scripts\") pod \"nova-cell0-db-create-kp7fb\" (UID: \"539b8651-7e63-416c-8428-83cd2e4b3813\") " pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.218775 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-config-data\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.218864 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlwpf\" (UniqueName: \"kubernetes.io/projected/539b8651-7e63-416c-8428-83cd2e4b3813-kube-api-access-mlwpf\") pod \"nova-cell0-db-create-kp7fb\" (UID: \"539b8651-7e63-416c-8428-83cd2e4b3813\") " pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.218948 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff27d27-5fa2-401b-95a2-c633aa025759-operator-scripts\") pod \"nova-api-db-create-2kcns\" (UID: \"7ff27d27-5fa2-401b-95a2-c633aa025759\") " pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.219096 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.219320 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpznz\" (UniqueName: \"kubernetes.io/projected/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-kube-api-access-lpznz\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.219428 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-scripts\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.219461 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-run-httpd\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.219756 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-log-httpd\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.220583 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff27d27-5fa2-401b-95a2-c633aa025759-operator-scripts\") pod \"nova-api-db-create-2kcns\" (UID: \"7ff27d27-5fa2-401b-95a2-c633aa025759\") " pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.229752 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-config-data\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.237454 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-scripts\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.238013 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.243908 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-9rq88"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.245301 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.262157 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.272256 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-96ff-account-create-update-mt94j"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.274257 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.276493 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpznz\" (UniqueName: \"kubernetes.io/projected/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-kube-api-access-lpznz\") pod \"ceilometer-0\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.278939 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.322130 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgqdq\" (UniqueName: \"kubernetes.io/projected/7ff27d27-5fa2-401b-95a2-c633aa025759-kube-api-access-fgqdq\") pod \"nova-api-db-create-2kcns\" (UID: \"7ff27d27-5fa2-401b-95a2-c633aa025759\") " pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.329983 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-operator-scripts\") pod \"nova-cell1-db-create-9rq88\" (UID: \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\") " pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.330077 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm5lp\" (UniqueName: \"kubernetes.io/projected/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-kube-api-access-hm5lp\") pod \"nova-cell1-db-create-9rq88\" (UID: \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\") " pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.330118 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9rq88"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.330321 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/539b8651-7e63-416c-8428-83cd2e4b3813-operator-scripts\") pod \"nova-cell0-db-create-kp7fb\" (UID: \"539b8651-7e63-416c-8428-83cd2e4b3813\") " pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.330465 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlwpf\" (UniqueName: \"kubernetes.io/projected/539b8651-7e63-416c-8428-83cd2e4b3813-kube-api-access-mlwpf\") pod \"nova-cell0-db-create-kp7fb\" (UID: \"539b8651-7e63-416c-8428-83cd2e4b3813\") " pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.331779 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/539b8651-7e63-416c-8428-83cd2e4b3813-operator-scripts\") pod \"nova-cell0-db-create-kp7fb\" (UID: \"539b8651-7e63-416c-8428-83cd2e4b3813\") " pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.332596 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.354317 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-96ff-account-create-update-mt94j"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.368430 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlwpf\" (UniqueName: \"kubernetes.io/projected/539b8651-7e63-416c-8428-83cd2e4b3813-kube-api-access-mlwpf\") pod \"nova-cell0-db-create-kp7fb\" (UID: \"539b8651-7e63-416c-8428-83cd2e4b3813\") " pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.411517 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.440805 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-operator-scripts\") pod \"nova-cell1-db-create-9rq88\" (UID: \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\") " pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.440860 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm5lp\" (UniqueName: \"kubernetes.io/projected/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-kube-api-access-hm5lp\") pod \"nova-cell1-db-create-9rq88\" (UID: \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\") " pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.440926 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-operator-scripts\") pod \"nova-api-96ff-account-create-update-mt94j\" (UID: \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\") " pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.440964 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9tqs\" (UniqueName: \"kubernetes.io/projected/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-kube-api-access-q9tqs\") pod \"nova-api-96ff-account-create-update-mt94j\" (UID: \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\") " pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.442069 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-operator-scripts\") pod \"nova-cell1-db-create-9rq88\" (UID: \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\") " pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.448120 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.462991 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm5lp\" (UniqueName: \"kubernetes.io/projected/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-kube-api-access-hm5lp\") pod \"nova-cell1-db-create-9rq88\" (UID: \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\") " pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.531170 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-b5d2-account-create-update-wr7tj"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.532676 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.540032 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.543147 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-operator-scripts\") pod \"nova-api-96ff-account-create-update-mt94j\" (UID: \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\") " pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.543206 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9tqs\" (UniqueName: \"kubernetes.io/projected/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-kube-api-access-q9tqs\") pod \"nova-api-96ff-account-create-update-mt94j\" (UID: \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\") " pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.544275 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-operator-scripts\") pod \"nova-api-96ff-account-create-update-mt94j\" (UID: \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\") " pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.549458 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b5d2-account-create-update-wr7tj"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.577517 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9tqs\" (UniqueName: \"kubernetes.io/projected/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-kube-api-access-q9tqs\") pod \"nova-api-96ff-account-create-update-mt94j\" (UID: \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\") " pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.656891 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d287bf0-5630-48c1-89c7-2c3682991e6d-operator-scripts\") pod \"nova-cell0-b5d2-account-create-update-wr7tj\" (UID: \"3d287bf0-5630-48c1-89c7-2c3682991e6d\") " pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.659094 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xvcw\" (UniqueName: \"kubernetes.io/projected/3d287bf0-5630-48c1-89c7-2c3682991e6d-kube-api-access-7xvcw\") pod \"nova-cell0-b5d2-account-create-update-wr7tj\" (UID: \"3d287bf0-5630-48c1-89c7-2c3682991e6d\") " pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.715971 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-00ab-account-create-update-9bqrf"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.739438 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.756229 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.757324 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.761116 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-00ab-account-create-update-9bqrf"] Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.772383 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.781703 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d287bf0-5630-48c1-89c7-2c3682991e6d-operator-scripts\") pod \"nova-cell0-b5d2-account-create-update-wr7tj\" (UID: \"3d287bf0-5630-48c1-89c7-2c3682991e6d\") " pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.781877 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-operator-scripts\") pod \"nova-cell1-00ab-account-create-update-9bqrf\" (UID: \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\") " pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.781963 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xvcw\" (UniqueName: \"kubernetes.io/projected/3d287bf0-5630-48c1-89c7-2c3682991e6d-kube-api-access-7xvcw\") pod \"nova-cell0-b5d2-account-create-update-wr7tj\" (UID: \"3d287bf0-5630-48c1-89c7-2c3682991e6d\") " pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.782161 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5swsw\" (UniqueName: \"kubernetes.io/projected/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-kube-api-access-5swsw\") pod \"nova-cell1-00ab-account-create-update-9bqrf\" (UID: \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\") " pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.783450 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d287bf0-5630-48c1-89c7-2c3682991e6d-operator-scripts\") pod \"nova-cell0-b5d2-account-create-update-wr7tj\" (UID: \"3d287bf0-5630-48c1-89c7-2c3682991e6d\") " pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.834835 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xvcw\" (UniqueName: \"kubernetes.io/projected/3d287bf0-5630-48c1-89c7-2c3682991e6d-kube-api-access-7xvcw\") pod \"nova-cell0-b5d2-account-create-update-wr7tj\" (UID: \"3d287bf0-5630-48c1-89c7-2c3682991e6d\") " pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.853974 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.893258 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-operator-scripts\") pod \"nova-cell1-00ab-account-create-update-9bqrf\" (UID: \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\") " pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.893564 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5swsw\" (UniqueName: \"kubernetes.io/projected/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-kube-api-access-5swsw\") pod \"nova-cell1-00ab-account-create-update-9bqrf\" (UID: \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\") " pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.894819 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-operator-scripts\") pod \"nova-cell1-00ab-account-create-update-9bqrf\" (UID: \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\") " pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:52 crc kubenswrapper[4573]: I1203 09:02:52.911752 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5swsw\" (UniqueName: \"kubernetes.io/projected/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-kube-api-access-5swsw\") pod \"nova-cell1-00ab-account-create-update-9bqrf\" (UID: \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\") " pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:53 crc kubenswrapper[4573]: I1203 09:02:53.103859 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:53 crc kubenswrapper[4573]: I1203 09:02:53.314234 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:53 crc kubenswrapper[4573]: I1203 09:02:53.589272 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2kcns"] Dec 03 09:02:53 crc kubenswrapper[4573]: I1203 09:02:53.670134 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-kp7fb"] Dec 03 09:02:53 crc kubenswrapper[4573]: W1203 09:02:53.678894 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod539b8651_7e63_416c_8428_83cd2e4b3813.slice/crio-5a897d2f023cdde8d8ba5bdcc1ce8deaf7ec1f398c31c225afcf7c64d32ff164 WatchSource:0}: Error finding container 5a897d2f023cdde8d8ba5bdcc1ce8deaf7ec1f398c31c225afcf7c64d32ff164: Status 404 returned error can't find the container with id 5a897d2f023cdde8d8ba5bdcc1ce8deaf7ec1f398c31c225afcf7c64d32ff164 Dec 03 09:02:53 crc kubenswrapper[4573]: I1203 09:02:53.751556 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kp7fb" event={"ID":"539b8651-7e63-416c-8428-83cd2e4b3813","Type":"ContainerStarted","Data":"5a897d2f023cdde8d8ba5bdcc1ce8deaf7ec1f398c31c225afcf7c64d32ff164"} Dec 03 09:02:53 crc kubenswrapper[4573]: I1203 09:02:53.754193 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerStarted","Data":"ea6332ee68364073a374be440c9f035e93483bbfae7b6409bcaebffca8004e4d"} Dec 03 09:02:53 crc kubenswrapper[4573]: I1203 09:02:53.756905 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2kcns" event={"ID":"7ff27d27-5fa2-401b-95a2-c633aa025759","Type":"ContainerStarted","Data":"2822f046eb71ae5b5dc81a768ef5786f34b79a8d0f508ac4daaee5cc2ade0518"} Dec 03 09:02:53 crc kubenswrapper[4573]: I1203 09:02:53.961398 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9rq88"] Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.020827 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-00ab-account-create-update-9bqrf"] Dec 03 09:02:54 crc kubenswrapper[4573]: W1203 09:02:54.056464 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e2ebf6a_1287_48f4_a13d_4a31ae5c10e0.slice/crio-d7c9db62541260fda21aa49a3666af120769024f92dbed0c2d6fb6dc08e8356b WatchSource:0}: Error finding container d7c9db62541260fda21aa49a3666af120769024f92dbed0c2d6fb6dc08e8356b: Status 404 returned error can't find the container with id d7c9db62541260fda21aa49a3666af120769024f92dbed0c2d6fb6dc08e8356b Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.084954 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-96ff-account-create-update-mt94j"] Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.108112 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b5d2-account-create-update-wr7tj"] Dec 03 09:02:54 crc kubenswrapper[4573]: W1203 09:02:54.155799 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f85ea37_8ca3_4b36_9c98_d1b2a51d108e.slice/crio-02ac40aa5d382a59484ec45b2b137bef0fd78f93ca45c78b9da83241be1a31ec WatchSource:0}: Error finding container 02ac40aa5d382a59484ec45b2b137bef0fd78f93ca45c78b9da83241be1a31ec: Status 404 returned error can't find the container with id 02ac40aa5d382a59484ec45b2b137bef0fd78f93ca45c78b9da83241be1a31ec Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.775708 4573 generic.go:334] "Generic (PLEG): container finished" podID="7ff27d27-5fa2-401b-95a2-c633aa025759" containerID="720b8a43dea71e416a2c2843eaa0945696e842d829ec3f58a6988e63bf047368" exitCode=0 Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.775999 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2kcns" event={"ID":"7ff27d27-5fa2-401b-95a2-c633aa025759","Type":"ContainerDied","Data":"720b8a43dea71e416a2c2843eaa0945696e842d829ec3f58a6988e63bf047368"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.786896 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" event={"ID":"3d287bf0-5630-48c1-89c7-2c3682991e6d","Type":"ContainerStarted","Data":"a0f77cc7444581881aaea554f7a0c8d6b00b893c572ea1029e9064079cae3832"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.786948 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" event={"ID":"3d287bf0-5630-48c1-89c7-2c3682991e6d","Type":"ContainerStarted","Data":"7efa233aa17afb7dc2bef806ff7f2b8f744fd5764ce39f0561e746bfb78cff86"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.800956 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-96ff-account-create-update-mt94j" event={"ID":"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e","Type":"ContainerStarted","Data":"fb0eb86bfd54149afc35d20883842b27700bb65d2b26f1a86893e9f54c787f13"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.801022 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-96ff-account-create-update-mt94j" event={"ID":"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e","Type":"ContainerStarted","Data":"02ac40aa5d382a59484ec45b2b137bef0fd78f93ca45c78b9da83241be1a31ec"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.824765 4573 generic.go:334] "Generic (PLEG): container finished" podID="539b8651-7e63-416c-8428-83cd2e4b3813" containerID="4988230dc7c3ce4c513117a7eb23941921f99b876e24ee199038a3618c4f4373" exitCode=0 Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.824909 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kp7fb" event={"ID":"539b8651-7e63-416c-8428-83cd2e4b3813","Type":"ContainerDied","Data":"4988230dc7c3ce4c513117a7eb23941921f99b876e24ee199038a3618c4f4373"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.847882 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-96ff-account-create-update-mt94j" podStartSLOduration=2.84785144 podStartE2EDuration="2.84785144s" podCreationTimestamp="2025-12-03 09:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:02:54.822341589 +0000 UTC m=+1495.390720858" watchObservedRunningTime="2025-12-03 09:02:54.84785144 +0000 UTC m=+1495.416230699" Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.855828 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" podStartSLOduration=2.855804689 podStartE2EDuration="2.855804689s" podCreationTimestamp="2025-12-03 09:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:02:54.849783223 +0000 UTC m=+1495.418162493" watchObservedRunningTime="2025-12-03 09:02:54.855804689 +0000 UTC m=+1495.424183948" Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.888486 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" event={"ID":"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0","Type":"ContainerStarted","Data":"6b56c6795993e4802102a503d6b4bdb93cd5a0adb04f492a4a89db636958a5f9"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.888528 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" event={"ID":"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0","Type":"ContainerStarted","Data":"d7c9db62541260fda21aa49a3666af120769024f92dbed0c2d6fb6dc08e8356b"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.898448 4573 generic.go:334] "Generic (PLEG): container finished" podID="7ad404c4-6b5d-4d8b-8298-efce4c5d54bb" containerID="340b52b445e91622c47165bd2cbbdd1cffe6096bbce9a1af515a640646a9312a" exitCode=0 Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.898551 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9rq88" event={"ID":"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb","Type":"ContainerDied","Data":"340b52b445e91622c47165bd2cbbdd1cffe6096bbce9a1af515a640646a9312a"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.898590 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9rq88" event={"ID":"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb","Type":"ContainerStarted","Data":"26508788deac5bc428a1e80ed5fe0f0687c0324ca040837b93af230b9d90ad31"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.905541 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerStarted","Data":"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1"} Dec 03 09:02:54 crc kubenswrapper[4573]: I1203 09:02:54.930609 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" podStartSLOduration=2.930578164 podStartE2EDuration="2.930578164s" podCreationTimestamp="2025-12-03 09:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:02:54.918837601 +0000 UTC m=+1495.487216870" watchObservedRunningTime="2025-12-03 09:02:54.930578164 +0000 UTC m=+1495.498957433" Dec 03 09:02:55 crc kubenswrapper[4573]: I1203 09:02:55.551808 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:55 crc kubenswrapper[4573]: I1203 09:02:55.924361 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerStarted","Data":"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a"} Dec 03 09:02:55 crc kubenswrapper[4573]: I1203 09:02:55.929833 4573 generic.go:334] "Generic (PLEG): container finished" podID="3d287bf0-5630-48c1-89c7-2c3682991e6d" containerID="a0f77cc7444581881aaea554f7a0c8d6b00b893c572ea1029e9064079cae3832" exitCode=0 Dec 03 09:02:55 crc kubenswrapper[4573]: I1203 09:02:55.929964 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" event={"ID":"3d287bf0-5630-48c1-89c7-2c3682991e6d","Type":"ContainerDied","Data":"a0f77cc7444581881aaea554f7a0c8d6b00b893c572ea1029e9064079cae3832"} Dec 03 09:02:55 crc kubenswrapper[4573]: I1203 09:02:55.935901 4573 generic.go:334] "Generic (PLEG): container finished" podID="0f85ea37-8ca3-4b36-9c98-d1b2a51d108e" containerID="fb0eb86bfd54149afc35d20883842b27700bb65d2b26f1a86893e9f54c787f13" exitCode=0 Dec 03 09:02:55 crc kubenswrapper[4573]: I1203 09:02:55.935963 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-96ff-account-create-update-mt94j" event={"ID":"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e","Type":"ContainerDied","Data":"fb0eb86bfd54149afc35d20883842b27700bb65d2b26f1a86893e9f54c787f13"} Dec 03 09:02:55 crc kubenswrapper[4573]: I1203 09:02:55.940520 4573 generic.go:334] "Generic (PLEG): container finished" podID="5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0" containerID="6b56c6795993e4802102a503d6b4bdb93cd5a0adb04f492a4a89db636958a5f9" exitCode=0 Dec 03 09:02:55 crc kubenswrapper[4573]: I1203 09:02:55.940736 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" event={"ID":"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0","Type":"ContainerDied","Data":"6b56c6795993e4802102a503d6b4bdb93cd5a0adb04f492a4a89db636958a5f9"} Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.491762 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.608126 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff27d27-5fa2-401b-95a2-c633aa025759-operator-scripts\") pod \"7ff27d27-5fa2-401b-95a2-c633aa025759\" (UID: \"7ff27d27-5fa2-401b-95a2-c633aa025759\") " Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.608194 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgqdq\" (UniqueName: \"kubernetes.io/projected/7ff27d27-5fa2-401b-95a2-c633aa025759-kube-api-access-fgqdq\") pod \"7ff27d27-5fa2-401b-95a2-c633aa025759\" (UID: \"7ff27d27-5fa2-401b-95a2-c633aa025759\") " Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.609586 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ff27d27-5fa2-401b-95a2-c633aa025759-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ff27d27-5fa2-401b-95a2-c633aa025759" (UID: "7ff27d27-5fa2-401b-95a2-c633aa025759"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.615340 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff27d27-5fa2-401b-95a2-c633aa025759-kube-api-access-fgqdq" (OuterVolumeSpecName: "kube-api-access-fgqdq") pod "7ff27d27-5fa2-401b-95a2-c633aa025759" (UID: "7ff27d27-5fa2-401b-95a2-c633aa025759"). InnerVolumeSpecName "kube-api-access-fgqdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.665620 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.700137 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.725026 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm5lp\" (UniqueName: \"kubernetes.io/projected/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-kube-api-access-hm5lp\") pod \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\" (UID: \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\") " Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.725187 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlwpf\" (UniqueName: \"kubernetes.io/projected/539b8651-7e63-416c-8428-83cd2e4b3813-kube-api-access-mlwpf\") pod \"539b8651-7e63-416c-8428-83cd2e4b3813\" (UID: \"539b8651-7e63-416c-8428-83cd2e4b3813\") " Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.725367 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-operator-scripts\") pod \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\" (UID: \"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb\") " Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.725476 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/539b8651-7e63-416c-8428-83cd2e4b3813-operator-scripts\") pod \"539b8651-7e63-416c-8428-83cd2e4b3813\" (UID: \"539b8651-7e63-416c-8428-83cd2e4b3813\") " Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.726501 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ff27d27-5fa2-401b-95a2-c633aa025759-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.726532 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgqdq\" (UniqueName: \"kubernetes.io/projected/7ff27d27-5fa2-401b-95a2-c633aa025759-kube-api-access-fgqdq\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.727714 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/539b8651-7e63-416c-8428-83cd2e4b3813-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "539b8651-7e63-416c-8428-83cd2e4b3813" (UID: "539b8651-7e63-416c-8428-83cd2e4b3813"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.729593 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ad404c4-6b5d-4d8b-8298-efce4c5d54bb" (UID: "7ad404c4-6b5d-4d8b-8298-efce4c5d54bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.743334 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-kube-api-access-hm5lp" (OuterVolumeSpecName: "kube-api-access-hm5lp") pod "7ad404c4-6b5d-4d8b-8298-efce4c5d54bb" (UID: "7ad404c4-6b5d-4d8b-8298-efce4c5d54bb"). InnerVolumeSpecName "kube-api-access-hm5lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.743440 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/539b8651-7e63-416c-8428-83cd2e4b3813-kube-api-access-mlwpf" (OuterVolumeSpecName: "kube-api-access-mlwpf") pod "539b8651-7e63-416c-8428-83cd2e4b3813" (UID: "539b8651-7e63-416c-8428-83cd2e4b3813"). InnerVolumeSpecName "kube-api-access-mlwpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.828722 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlwpf\" (UniqueName: \"kubernetes.io/projected/539b8651-7e63-416c-8428-83cd2e4b3813-kube-api-access-mlwpf\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.828771 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.828781 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/539b8651-7e63-416c-8428-83cd2e4b3813-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.828791 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm5lp\" (UniqueName: \"kubernetes.io/projected/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb-kube-api-access-hm5lp\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.952148 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerStarted","Data":"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3"} Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.954069 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2kcns" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.954024 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2kcns" event={"ID":"7ff27d27-5fa2-401b-95a2-c633aa025759","Type":"ContainerDied","Data":"2822f046eb71ae5b5dc81a768ef5786f34b79a8d0f508ac4daaee5cc2ade0518"} Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.954372 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2822f046eb71ae5b5dc81a768ef5786f34b79a8d0f508ac4daaee5cc2ade0518" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.955894 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kp7fb" event={"ID":"539b8651-7e63-416c-8428-83cd2e4b3813","Type":"ContainerDied","Data":"5a897d2f023cdde8d8ba5bdcc1ce8deaf7ec1f398c31c225afcf7c64d32ff164"} Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.955965 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a897d2f023cdde8d8ba5bdcc1ce8deaf7ec1f398c31c225afcf7c64d32ff164" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.955919 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kp7fb" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.957246 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9rq88" Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.957669 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9rq88" event={"ID":"7ad404c4-6b5d-4d8b-8298-efce4c5d54bb","Type":"ContainerDied","Data":"26508788deac5bc428a1e80ed5fe0f0687c0324ca040837b93af230b9d90ad31"} Dec 03 09:02:56 crc kubenswrapper[4573]: I1203 09:02:56.957701 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26508788deac5bc428a1e80ed5fe0f0687c0324ca040837b93af230b9d90ad31" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.306825 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.348459 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d287bf0-5630-48c1-89c7-2c3682991e6d-operator-scripts\") pod \"3d287bf0-5630-48c1-89c7-2c3682991e6d\" (UID: \"3d287bf0-5630-48c1-89c7-2c3682991e6d\") " Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.348790 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xvcw\" (UniqueName: \"kubernetes.io/projected/3d287bf0-5630-48c1-89c7-2c3682991e6d-kube-api-access-7xvcw\") pod \"3d287bf0-5630-48c1-89c7-2c3682991e6d\" (UID: \"3d287bf0-5630-48c1-89c7-2c3682991e6d\") " Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.349329 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d287bf0-5630-48c1-89c7-2c3682991e6d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3d287bf0-5630-48c1-89c7-2c3682991e6d" (UID: "3d287bf0-5630-48c1-89c7-2c3682991e6d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.359401 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d287bf0-5630-48c1-89c7-2c3682991e6d-kube-api-access-7xvcw" (OuterVolumeSpecName: "kube-api-access-7xvcw") pod "3d287bf0-5630-48c1-89c7-2c3682991e6d" (UID: "3d287bf0-5630-48c1-89c7-2c3682991e6d"). InnerVolumeSpecName "kube-api-access-7xvcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.452379 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xvcw\" (UniqueName: \"kubernetes.io/projected/3d287bf0-5630-48c1-89c7-2c3682991e6d-kube-api-access-7xvcw\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.452463 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d287bf0-5630-48c1-89c7-2c3682991e6d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.641687 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.653419 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.665113 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-operator-scripts\") pod \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\" (UID: \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\") " Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.665653 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5swsw\" (UniqueName: \"kubernetes.io/projected/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-kube-api-access-5swsw\") pod \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\" (UID: \"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0\") " Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.665795 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0" (UID: "5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.666151 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.672369 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-kube-api-access-5swsw" (OuterVolumeSpecName: "kube-api-access-5swsw") pod "5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0" (UID: "5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0"). InnerVolumeSpecName "kube-api-access-5swsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.768710 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9tqs\" (UniqueName: \"kubernetes.io/projected/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-kube-api-access-q9tqs\") pod \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\" (UID: \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\") " Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.769732 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-operator-scripts\") pod \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\" (UID: \"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e\") " Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.770249 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5swsw\" (UniqueName: \"kubernetes.io/projected/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0-kube-api-access-5swsw\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.772886 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f85ea37-8ca3-4b36-9c98-d1b2a51d108e" (UID: "0f85ea37-8ca3-4b36-9c98-d1b2a51d108e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.777152 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-kube-api-access-q9tqs" (OuterVolumeSpecName: "kube-api-access-q9tqs") pod "0f85ea37-8ca3-4b36-9c98-d1b2a51d108e" (UID: "0f85ea37-8ca3-4b36-9c98-d1b2a51d108e"). InnerVolumeSpecName "kube-api-access-q9tqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.873471 4573 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.873531 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9tqs\" (UniqueName: \"kubernetes.io/projected/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e-kube-api-access-q9tqs\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.972935 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-96ff-account-create-update-mt94j" event={"ID":"0f85ea37-8ca3-4b36-9c98-d1b2a51d108e","Type":"ContainerDied","Data":"02ac40aa5d382a59484ec45b2b137bef0fd78f93ca45c78b9da83241be1a31ec"} Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.973002 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02ac40aa5d382a59484ec45b2b137bef0fd78f93ca45c78b9da83241be1a31ec" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.973148 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-96ff-account-create-update-mt94j" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.982588 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" event={"ID":"5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0","Type":"ContainerDied","Data":"d7c9db62541260fda21aa49a3666af120769024f92dbed0c2d6fb6dc08e8356b"} Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.982732 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7c9db62541260fda21aa49a3666af120769024f92dbed0c2d6fb6dc08e8356b" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.982598 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-00ab-account-create-update-9bqrf" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.990559 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerStarted","Data":"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d"} Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.990720 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.990773 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="ceilometer-central-agent" containerID="cri-o://7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1" gracePeriod=30 Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.990857 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="ceilometer-notification-agent" containerID="cri-o://dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a" gracePeriod=30 Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.990794 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="proxy-httpd" containerID="cri-o://b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d" gracePeriod=30 Dec 03 09:02:57 crc kubenswrapper[4573]: I1203 09:02:57.990964 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="sg-core" containerID="cri-o://bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3" gracePeriod=30 Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:57.998547 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" event={"ID":"3d287bf0-5630-48c1-89c7-2c3682991e6d","Type":"ContainerDied","Data":"7efa233aa17afb7dc2bef806ff7f2b8f744fd5764ce39f0561e746bfb78cff86"} Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:57.998587 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7efa233aa17afb7dc2bef806ff7f2b8f744fd5764ce39f0561e746bfb78cff86" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:57.998649 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b5d2-account-create-update-wr7tj" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.030445 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.058289558 podStartE2EDuration="7.030423735s" podCreationTimestamp="2025-12-03 09:02:51 +0000 UTC" firstStartedPulling="2025-12-03 09:02:53.405236779 +0000 UTC m=+1493.973616038" lastFinishedPulling="2025-12-03 09:02:57.377370956 +0000 UTC m=+1497.945750215" observedRunningTime="2025-12-03 09:02:58.022016904 +0000 UTC m=+1498.590396173" watchObservedRunningTime="2025-12-03 09:02:58.030423735 +0000 UTC m=+1498.598802994" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.783239 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.897239 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-scripts\") pod \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.897366 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-log-httpd\") pod \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.897390 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-config-data\") pod \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.897467 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-sg-core-conf-yaml\") pod \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.897512 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-run-httpd\") pod \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.897537 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpznz\" (UniqueName: \"kubernetes.io/projected/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-kube-api-access-lpznz\") pod \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.897564 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-combined-ca-bundle\") pod \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\" (UID: \"909a150d-dd82-48ae-ae9d-39f7aabd4ac8\") " Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.900271 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "909a150d-dd82-48ae-ae9d-39f7aabd4ac8" (UID: "909a150d-dd82-48ae-ae9d-39f7aabd4ac8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.900282 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "909a150d-dd82-48ae-ae9d-39f7aabd4ac8" (UID: "909a150d-dd82-48ae-ae9d-39f7aabd4ac8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.905039 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-scripts" (OuterVolumeSpecName: "scripts") pod "909a150d-dd82-48ae-ae9d-39f7aabd4ac8" (UID: "909a150d-dd82-48ae-ae9d-39f7aabd4ac8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.905134 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-kube-api-access-lpznz" (OuterVolumeSpecName: "kube-api-access-lpznz") pod "909a150d-dd82-48ae-ae9d-39f7aabd4ac8" (UID: "909a150d-dd82-48ae-ae9d-39f7aabd4ac8"). InnerVolumeSpecName "kube-api-access-lpznz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.941180 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "909a150d-dd82-48ae-ae9d-39f7aabd4ac8" (UID: "909a150d-dd82-48ae-ae9d-39f7aabd4ac8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:58 crc kubenswrapper[4573]: I1203 09:02:58.994546 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "909a150d-dd82-48ae-ae9d-39f7aabd4ac8" (UID: "909a150d-dd82-48ae-ae9d-39f7aabd4ac8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.001582 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.001642 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.001672 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.001699 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpznz\" (UniqueName: \"kubernetes.io/projected/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-kube-api-access-lpznz\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.001727 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.001854 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.012531 4573 generic.go:334] "Generic (PLEG): container finished" podID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerID="b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d" exitCode=0 Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.012646 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.012647 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerDied","Data":"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d"} Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.012736 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerDied","Data":"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3"} Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.012758 4573 scope.go:117] "RemoveContainer" containerID="b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.013596 4573 generic.go:334] "Generic (PLEG): container finished" podID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerID="bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3" exitCode=2 Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.013806 4573 generic.go:334] "Generic (PLEG): container finished" podID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerID="dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a" exitCode=0 Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.013835 4573 generic.go:334] "Generic (PLEG): container finished" podID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerID="7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1" exitCode=0 Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.013836 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerDied","Data":"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a"} Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.013892 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerDied","Data":"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1"} Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.013910 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"909a150d-dd82-48ae-ae9d-39f7aabd4ac8","Type":"ContainerDied","Data":"ea6332ee68364073a374be440c9f035e93483bbfae7b6409bcaebffca8004e4d"} Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.014577 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-config-data" (OuterVolumeSpecName: "config-data") pod "909a150d-dd82-48ae-ae9d-39f7aabd4ac8" (UID: "909a150d-dd82-48ae-ae9d-39f7aabd4ac8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.037640 4573 scope.go:117] "RemoveContainer" containerID="bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.063596 4573 scope.go:117] "RemoveContainer" containerID="dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.090540 4573 scope.go:117] "RemoveContainer" containerID="7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.103365 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909a150d-dd82-48ae-ae9d-39f7aabd4ac8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.110978 4573 scope.go:117] "RemoveContainer" containerID="b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.111402 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": container with ID starting with b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d not found: ID does not exist" containerID="b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.111535 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d"} err="failed to get container status \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": rpc error: code = NotFound desc = could not find container \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": container with ID starting with b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.111619 4573 scope.go:117] "RemoveContainer" containerID="bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.112149 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": container with ID starting with bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3 not found: ID does not exist" containerID="bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.112189 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3"} err="failed to get container status \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": rpc error: code = NotFound desc = could not find container \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": container with ID starting with bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3 not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.112216 4573 scope.go:117] "RemoveContainer" containerID="dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.114179 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": container with ID starting with dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a not found: ID does not exist" containerID="dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.114209 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a"} err="failed to get container status \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": rpc error: code = NotFound desc = could not find container \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": container with ID starting with dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.114226 4573 scope.go:117] "RemoveContainer" containerID="7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.114518 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": container with ID starting with 7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1 not found: ID does not exist" containerID="7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.114545 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1"} err="failed to get container status \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": rpc error: code = NotFound desc = could not find container \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": container with ID starting with 7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1 not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.114564 4573 scope.go:117] "RemoveContainer" containerID="b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.114918 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d"} err="failed to get container status \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": rpc error: code = NotFound desc = could not find container \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": container with ID starting with b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.114967 4573 scope.go:117] "RemoveContainer" containerID="bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.115431 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3"} err="failed to get container status \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": rpc error: code = NotFound desc = could not find container \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": container with ID starting with bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3 not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.115453 4573 scope.go:117] "RemoveContainer" containerID="dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.115733 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a"} err="failed to get container status \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": rpc error: code = NotFound desc = could not find container \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": container with ID starting with dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.115756 4573 scope.go:117] "RemoveContainer" containerID="7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.116029 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1"} err="failed to get container status \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": rpc error: code = NotFound desc = could not find container \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": container with ID starting with 7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1 not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.116064 4573 scope.go:117] "RemoveContainer" containerID="b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.116290 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d"} err="failed to get container status \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": rpc error: code = NotFound desc = could not find container \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": container with ID starting with b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.116309 4573 scope.go:117] "RemoveContainer" containerID="bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.116717 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3"} err="failed to get container status \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": rpc error: code = NotFound desc = could not find container \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": container with ID starting with bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3 not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.116766 4573 scope.go:117] "RemoveContainer" containerID="dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.117014 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a"} err="failed to get container status \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": rpc error: code = NotFound desc = could not find container \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": container with ID starting with dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.117038 4573 scope.go:117] "RemoveContainer" containerID="7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.117294 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1"} err="failed to get container status \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": rpc error: code = NotFound desc = could not find container \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": container with ID starting with 7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1 not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.117317 4573 scope.go:117] "RemoveContainer" containerID="b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.117771 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d"} err="failed to get container status \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": rpc error: code = NotFound desc = could not find container \"b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d\": container with ID starting with b29a323b55d15776831394c74160027d9bfd3f8cacdd327697a5604705488a2d not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.117805 4573 scope.go:117] "RemoveContainer" containerID="bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.118036 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3"} err="failed to get container status \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": rpc error: code = NotFound desc = could not find container \"bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3\": container with ID starting with bed7722176c784390c6f5a084c3ba36ea25bfea74c5a93148aa9bbab458c26c3 not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.118087 4573 scope.go:117] "RemoveContainer" containerID="dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.118579 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a"} err="failed to get container status \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": rpc error: code = NotFound desc = could not find container \"dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a\": container with ID starting with dc1cd3603d5466b9d6bf4f704a79983dca84b0a3726b5fc46180890dcd65209a not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.118600 4573 scope.go:117] "RemoveContainer" containerID="7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.118826 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1"} err="failed to get container status \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": rpc error: code = NotFound desc = could not find container \"7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1\": container with ID starting with 7b93d37f8358cddb1e955a63b0e7ed85c1ad0847243875519526a1b6ae7afdb1 not found: ID does not exist" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.431693 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.440783 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467138 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467596 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="proxy-httpd" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467622 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="proxy-httpd" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467638 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff27d27-5fa2-401b-95a2-c633aa025759" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467647 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff27d27-5fa2-401b-95a2-c633aa025759" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467677 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="ceilometer-notification-agent" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467689 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="ceilometer-notification-agent" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467701 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f85ea37-8ca3-4b36-9c98-d1b2a51d108e" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467709 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f85ea37-8ca3-4b36-9c98-d1b2a51d108e" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467723 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539b8651-7e63-416c-8428-83cd2e4b3813" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467732 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="539b8651-7e63-416c-8428-83cd2e4b3813" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467749 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d287bf0-5630-48c1-89c7-2c3682991e6d" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467757 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d287bf0-5630-48c1-89c7-2c3682991e6d" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467771 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad404c4-6b5d-4d8b-8298-efce4c5d54bb" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467779 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad404c4-6b5d-4d8b-8298-efce4c5d54bb" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467800 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="sg-core" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467810 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="sg-core" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467825 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="ceilometer-central-agent" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467833 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="ceilometer-central-agent" Dec 03 09:02:59 crc kubenswrapper[4573]: E1203 09:02:59.467852 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.467860 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468061 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468077 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="proxy-httpd" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468088 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="539b8651-7e63-416c-8428-83cd2e4b3813" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468098 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d287bf0-5630-48c1-89c7-2c3682991e6d" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468111 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff27d27-5fa2-401b-95a2-c633aa025759" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468119 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f85ea37-8ca3-4b36-9c98-d1b2a51d108e" containerName="mariadb-account-create-update" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468129 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="ceilometer-notification-agent" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468140 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="ceilometer-central-agent" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468148 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" containerName="sg-core" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.468160 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad404c4-6b5d-4d8b-8298-efce4c5d54bb" containerName="mariadb-database-create" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.470348 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.477747 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.479157 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.508603 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.616398 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-log-httpd\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.617138 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.617315 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kx6b\" (UniqueName: \"kubernetes.io/projected/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-kube-api-access-5kx6b\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.617433 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.617607 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-config-data\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.617750 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-scripts\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.617831 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-run-httpd\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.720109 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-scripts\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.720170 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-run-httpd\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.720257 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-log-httpd\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.720298 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.720333 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kx6b\" (UniqueName: \"kubernetes.io/projected/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-kube-api-access-5kx6b\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.720372 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.720437 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-config-data\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.722119 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-run-httpd\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.722703 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-log-httpd\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.726997 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.727847 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-config-data\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.729505 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.731306 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-scripts\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.745966 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kx6b\" (UniqueName: \"kubernetes.io/projected/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-kube-api-access-5kx6b\") pod \"ceilometer-0\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " pod="openstack/ceilometer-0" Dec 03 09:02:59 crc kubenswrapper[4573]: I1203 09:02:59.786164 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:03:00 crc kubenswrapper[4573]: I1203 09:03:00.054292 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="909a150d-dd82-48ae-ae9d-39f7aabd4ac8" path="/var/lib/kubelet/pods/909a150d-dd82-48ae-ae9d-39f7aabd4ac8/volumes" Dec 03 09:03:00 crc kubenswrapper[4573]: I1203 09:03:00.318205 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:03:01 crc kubenswrapper[4573]: I1203 09:03:01.060251 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerStarted","Data":"3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb"} Dec 03 09:03:01 crc kubenswrapper[4573]: I1203 09:03:01.061541 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerStarted","Data":"b1f26680c0d7a9356a65bdc1d4c46415f0c35f8648b53f5fb96dd85283bbfdfb"} Dec 03 09:03:02 crc kubenswrapper[4573]: I1203 09:03:02.112485 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerStarted","Data":"28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba"} Dec 03 09:03:02 crc kubenswrapper[4573]: I1203 09:03:02.996693 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-22zjj"] Dec 03 09:03:02 crc kubenswrapper[4573]: I1203 09:03:02.998901 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.005565 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.005820 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.006077 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-7b8bg" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.019034 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-22zjj"] Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.097542 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.097652 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q77rd\" (UniqueName: \"kubernetes.io/projected/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-kube-api-access-q77rd\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.097687 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-scripts\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.097745 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-config-data\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.148984 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerStarted","Data":"900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d"} Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.201533 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.201618 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q77rd\" (UniqueName: \"kubernetes.io/projected/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-kube-api-access-q77rd\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.201650 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-scripts\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.201678 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-config-data\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.213194 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-config-data\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.213323 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.217558 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-scripts\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.221556 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q77rd\" (UniqueName: \"kubernetes.io/projected/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-kube-api-access-q77rd\") pod \"nova-cell0-conductor-db-sync-22zjj\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.325925 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:03 crc kubenswrapper[4573]: I1203 09:03:03.960761 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-22zjj"] Dec 03 09:03:03 crc kubenswrapper[4573]: W1203 09:03:03.961291 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b4e8f12_4f4b_4ae7_a18b_2fe2dd9427cc.slice/crio-fe8aebc5e62fb36a8590bc5c4ad955205a3a78e0a3585c1dbccf4f064e236a4e WatchSource:0}: Error finding container fe8aebc5e62fb36a8590bc5c4ad955205a3a78e0a3585c1dbccf4f064e236a4e: Status 404 returned error can't find the container with id fe8aebc5e62fb36a8590bc5c4ad955205a3a78e0a3585c1dbccf4f064e236a4e Dec 03 09:03:04 crc kubenswrapper[4573]: I1203 09:03:04.164126 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-22zjj" event={"ID":"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc","Type":"ContainerStarted","Data":"fe8aebc5e62fb36a8590bc5c4ad955205a3a78e0a3585c1dbccf4f064e236a4e"} Dec 03 09:03:04 crc kubenswrapper[4573]: I1203 09:03:04.184678 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerStarted","Data":"e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244"} Dec 03 09:03:04 crc kubenswrapper[4573]: I1203 09:03:04.186335 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:03:04 crc kubenswrapper[4573]: I1203 09:03:04.227503 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.175672315 podStartE2EDuration="5.227476566s" podCreationTimestamp="2025-12-03 09:02:59 +0000 UTC" firstStartedPulling="2025-12-03 09:03:00.335123952 +0000 UTC m=+1500.903503201" lastFinishedPulling="2025-12-03 09:03:03.386928193 +0000 UTC m=+1503.955307452" observedRunningTime="2025-12-03 09:03:04.212620598 +0000 UTC m=+1504.780999867" watchObservedRunningTime="2025-12-03 09:03:04.227476566 +0000 UTC m=+1504.795855825" Dec 03 09:03:07 crc kubenswrapper[4573]: I1203 09:03:07.367467 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:03:07 crc kubenswrapper[4573]: I1203 09:03:07.368267 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="ceilometer-central-agent" containerID="cri-o://3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb" gracePeriod=30 Dec 03 09:03:07 crc kubenswrapper[4573]: I1203 09:03:07.368780 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="proxy-httpd" containerID="cri-o://e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244" gracePeriod=30 Dec 03 09:03:07 crc kubenswrapper[4573]: I1203 09:03:07.368822 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="sg-core" containerID="cri-o://900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d" gracePeriod=30 Dec 03 09:03:07 crc kubenswrapper[4573]: I1203 09:03:07.368853 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="ceilometer-notification-agent" containerID="cri-o://28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba" gracePeriod=30 Dec 03 09:03:08 crc kubenswrapper[4573]: I1203 09:03:08.245948 4573 generic.go:334] "Generic (PLEG): container finished" podID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerID="e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244" exitCode=0 Dec 03 09:03:08 crc kubenswrapper[4573]: I1203 09:03:08.246428 4573 generic.go:334] "Generic (PLEG): container finished" podID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerID="900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d" exitCode=2 Dec 03 09:03:08 crc kubenswrapper[4573]: I1203 09:03:08.246439 4573 generic.go:334] "Generic (PLEG): container finished" podID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerID="28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba" exitCode=0 Dec 03 09:03:08 crc kubenswrapper[4573]: I1203 09:03:08.246374 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerDied","Data":"e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244"} Dec 03 09:03:08 crc kubenswrapper[4573]: I1203 09:03:08.246482 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerDied","Data":"900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d"} Dec 03 09:03:08 crc kubenswrapper[4573]: I1203 09:03:08.246497 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerDied","Data":"28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba"} Dec 03 09:03:11 crc kubenswrapper[4573]: I1203 09:03:11.284384 4573 generic.go:334] "Generic (PLEG): container finished" podID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerID="eee5245aaff562e364c268a895dd3a53869d1cf3cec162d72a6fbfa629898941" exitCode=137 Dec 03 09:03:11 crc kubenswrapper[4573]: I1203 09:03:11.284703 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66f99c579b-gh7sh" event={"ID":"b78e2453-adc3-4139-97c5-9f8a977df1ee","Type":"ContainerDied","Data":"eee5245aaff562e364c268a895dd3a53869d1cf3cec162d72a6fbfa629898941"} Dec 03 09:03:11 crc kubenswrapper[4573]: I1203 09:03:11.285156 4573 scope.go:117] "RemoveContainer" containerID="afd0f349dd0be138a212ad8426fb19507ce3fcb56dbe6dbd953d89456cd91458" Dec 03 09:03:11 crc kubenswrapper[4573]: I1203 09:03:11.289016 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerID="b8aa7ba7c95bd0a8692253b9b93fecbaa60bed28d93ea2c5355468e1a8ef3047" exitCode=137 Dec 03 09:03:11 crc kubenswrapper[4573]: I1203 09:03:11.289079 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerDied","Data":"b8aa7ba7c95bd0a8692253b9b93fecbaa60bed28d93ea2c5355468e1a8ef3047"} Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.240460 4573 scope.go:117] "RemoveContainer" containerID="875e1f86deecceeb9c350e1909663db37b4f6f0a82cff5c1165cad17d9cdc1df" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.730259 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.782651 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-scripts\") pod \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.782745 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-config-data\") pod \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.782781 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-sg-core-conf-yaml\") pod \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.782801 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-combined-ca-bundle\") pod \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.782833 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-run-httpd\") pod \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.782895 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kx6b\" (UniqueName: \"kubernetes.io/projected/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-kube-api-access-5kx6b\") pod \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.782937 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-log-httpd\") pod \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\" (UID: \"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975\") " Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.783510 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" (UID: "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.783730 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" (UID: "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.788275 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-scripts" (OuterVolumeSpecName: "scripts") pod "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" (UID: "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.805341 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-kube-api-access-5kx6b" (OuterVolumeSpecName: "kube-api-access-5kx6b") pod "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" (UID: "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975"). InnerVolumeSpecName "kube-api-access-5kx6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.837231 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" (UID: "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.885443 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.885698 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.885768 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kx6b\" (UniqueName: \"kubernetes.io/projected/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-kube-api-access-5kx6b\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.885834 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.885896 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.907901 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-config-data" (OuterVolumeSpecName: "config-data") pod "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" (UID: "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.910865 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" (UID: "8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.988443 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:13 crc kubenswrapper[4573]: I1203 09:03:13.988854 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.334351 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerStarted","Data":"d0c803805d4bd3808e5ee89977579fff41f3ca9da62cdfa598f77cba0d2c5e22"} Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.345469 4573 generic.go:334] "Generic (PLEG): container finished" podID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerID="3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb" exitCode=0 Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.345972 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.347818 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerDied","Data":"3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb"} Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.348038 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975","Type":"ContainerDied","Data":"b1f26680c0d7a9356a65bdc1d4c46415f0c35f8648b53f5fb96dd85283bbfdfb"} Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.348218 4573 scope.go:117] "RemoveContainer" containerID="e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.352370 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-22zjj" event={"ID":"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc","Type":"ContainerStarted","Data":"e5abf83ae2290120529b2027ef11708338276a01b6082bda89e2b334c973e1ac"} Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.358690 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66f99c579b-gh7sh" event={"ID":"b78e2453-adc3-4139-97c5-9f8a977df1ee","Type":"ContainerStarted","Data":"7175eece615d27868fe5331b0c6561400b27029b5d362c3ba9e69d5252c5bce4"} Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.382817 4573 scope.go:117] "RemoveContainer" containerID="900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.408289 4573 scope.go:117] "RemoveContainer" containerID="28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.436825 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-22zjj" podStartSLOduration=3.159041651 podStartE2EDuration="12.436782935s" podCreationTimestamp="2025-12-03 09:03:02 +0000 UTC" firstStartedPulling="2025-12-03 09:03:03.963960123 +0000 UTC m=+1504.532339382" lastFinishedPulling="2025-12-03 09:03:13.241701407 +0000 UTC m=+1513.810080666" observedRunningTime="2025-12-03 09:03:14.396017404 +0000 UTC m=+1514.964396693" watchObservedRunningTime="2025-12-03 09:03:14.436782935 +0000 UTC m=+1515.005162204" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.465602 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.471374 4573 scope.go:117] "RemoveContainer" containerID="3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.483133 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.506355 4573 scope.go:117] "RemoveContainer" containerID="e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244" Dec 03 09:03:14 crc kubenswrapper[4573]: E1203 09:03:14.514426 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244\": container with ID starting with e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244 not found: ID does not exist" containerID="e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.514494 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244"} err="failed to get container status \"e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244\": rpc error: code = NotFound desc = could not find container \"e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244\": container with ID starting with e234a2d6805e18958a555284ed70f0fd4e12ddb9761a33804ac42568b01ab244 not found: ID does not exist" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.514522 4573 scope.go:117] "RemoveContainer" containerID="900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d" Dec 03 09:03:14 crc kubenswrapper[4573]: E1203 09:03:14.516425 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d\": container with ID starting with 900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d not found: ID does not exist" containerID="900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.516544 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d"} err="failed to get container status \"900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d\": rpc error: code = NotFound desc = could not find container \"900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d\": container with ID starting with 900f7ae71adce31fe8d47d411367308daf796433db9c91c9bc97fcfea5a7c63d not found: ID does not exist" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.516595 4573 scope.go:117] "RemoveContainer" containerID="28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba" Dec 03 09:03:14 crc kubenswrapper[4573]: E1203 09:03:14.528084 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba\": container with ID starting with 28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba not found: ID does not exist" containerID="28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.528147 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba"} err="failed to get container status \"28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba\": rpc error: code = NotFound desc = could not find container \"28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba\": container with ID starting with 28e81e3482569ddb6e2e5b6b5c6ba07c69f60c7e87e4ad1d2b660bceb89473ba not found: ID does not exist" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.528177 4573 scope.go:117] "RemoveContainer" containerID="3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb" Dec 03 09:03:14 crc kubenswrapper[4573]: E1203 09:03:14.536371 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb\": container with ID starting with 3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb not found: ID does not exist" containerID="3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.536436 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb"} err="failed to get container status \"3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb\": rpc error: code = NotFound desc = could not find container \"3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb\": container with ID starting with 3daba2f8c171f210b766a8bbd1d1ad5b30578414baed1a033ccd2f6dc69e4afb not found: ID does not exist" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.568618 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:03:14 crc kubenswrapper[4573]: E1203 09:03:14.569828 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="sg-core" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.569941 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="sg-core" Dec 03 09:03:14 crc kubenswrapper[4573]: E1203 09:03:14.570084 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="proxy-httpd" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.570161 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="proxy-httpd" Dec 03 09:03:14 crc kubenswrapper[4573]: E1203 09:03:14.570270 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="ceilometer-central-agent" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.570378 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="ceilometer-central-agent" Dec 03 09:03:14 crc kubenswrapper[4573]: E1203 09:03:14.570466 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="ceilometer-notification-agent" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.570533 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="ceilometer-notification-agent" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.571123 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="proxy-httpd" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.571233 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="ceilometer-notification-agent" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.571304 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="sg-core" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.571386 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" containerName="ceilometer-central-agent" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.575769 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.581113 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.581913 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.597447 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.720462 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-run-httpd\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.720718 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-scripts\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.720916 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw6f9\" (UniqueName: \"kubernetes.io/projected/441a63f9-caa0-4f25-8def-194be8fd64a6-kube-api-access-jw6f9\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.721016 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-config-data\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.721186 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-log-httpd\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.721354 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.721690 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.825470 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-log-httpd\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.825815 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.826070 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-log-httpd\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.827303 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.827446 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-run-httpd\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.827561 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-scripts\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.827813 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw6f9\" (UniqueName: \"kubernetes.io/projected/441a63f9-caa0-4f25-8def-194be8fd64a6-kube-api-access-jw6f9\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.827952 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-config-data\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.827964 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-run-httpd\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.850381 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.853604 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-scripts\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.859906 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.866390 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-config-data\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.868449 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw6f9\" (UniqueName: \"kubernetes.io/projected/441a63f9-caa0-4f25-8def-194be8fd64a6-kube-api-access-jw6f9\") pod \"ceilometer-0\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " pod="openstack/ceilometer-0" Dec 03 09:03:14 crc kubenswrapper[4573]: I1203 09:03:14.914850 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:03:15 crc kubenswrapper[4573]: I1203 09:03:15.450358 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:03:16 crc kubenswrapper[4573]: I1203 09:03:16.048717 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975" path="/var/lib/kubelet/pods/8d2c1b52-ab3b-4dc0-ac54-0076ec8e6975/volumes" Dec 03 09:03:16 crc kubenswrapper[4573]: I1203 09:03:16.389562 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerStarted","Data":"b5ebe341f35a6457c3af554054a07d18d857a6e65d782fe67d07eea441859d27"} Dec 03 09:03:16 crc kubenswrapper[4573]: I1203 09:03:16.389623 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerStarted","Data":"faf7b08b1d15f46454874a7248255627e413768aaa00da2e111f106fb98308a5"} Dec 03 09:03:18 crc kubenswrapper[4573]: I1203 09:03:18.431299 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerStarted","Data":"0fee465fc39dc7de24f61324f3f6b461424cc50d89e57c0b496e59334bd9b664"} Dec 03 09:03:18 crc kubenswrapper[4573]: I1203 09:03:18.431834 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerStarted","Data":"25248762cbd521ef8ef4803fd45cc8694960f9640ec665a520f524a5b23ada9c"} Dec 03 09:03:20 crc kubenswrapper[4573]: I1203 09:03:20.919093 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:03:20 crc kubenswrapper[4573]: I1203 09:03:20.920684 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:03:21 crc kubenswrapper[4573]: I1203 09:03:21.070698 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:03:21 crc kubenswrapper[4573]: I1203 09:03:21.070764 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:03:21 crc kubenswrapper[4573]: I1203 09:03:21.472342 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerStarted","Data":"be7b5ba9dcb78e51473052228d392d38368303937453aaf4de98c0c173805179"} Dec 03 09:03:21 crc kubenswrapper[4573]: I1203 09:03:21.472843 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:03:21 crc kubenswrapper[4573]: I1203 09:03:21.493976 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.493703251 podStartE2EDuration="7.493951777s" podCreationTimestamp="2025-12-03 09:03:14 +0000 UTC" firstStartedPulling="2025-12-03 09:03:15.455486764 +0000 UTC m=+1516.023866043" lastFinishedPulling="2025-12-03 09:03:20.45573531 +0000 UTC m=+1521.024114569" observedRunningTime="2025-12-03 09:03:21.491180931 +0000 UTC m=+1522.059560210" watchObservedRunningTime="2025-12-03 09:03:21.493951777 +0000 UTC m=+1522.062331036" Dec 03 09:03:29 crc kubenswrapper[4573]: I1203 09:03:29.553606 4573 generic.go:334] "Generic (PLEG): container finished" podID="4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" containerID="e5abf83ae2290120529b2027ef11708338276a01b6082bda89e2b334c973e1ac" exitCode=0 Dec 03 09:03:29 crc kubenswrapper[4573]: I1203 09:03:29.553713 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-22zjj" event={"ID":"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc","Type":"ContainerDied","Data":"e5abf83ae2290120529b2027ef11708338276a01b6082bda89e2b334c973e1ac"} Dec 03 09:03:30 crc kubenswrapper[4573]: I1203 09:03:30.920348 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 09:03:30 crc kubenswrapper[4573]: I1203 09:03:30.990306 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.041506 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-scripts\") pod \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.041566 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-combined-ca-bundle\") pod \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.041614 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q77rd\" (UniqueName: \"kubernetes.io/projected/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-kube-api-access-q77rd\") pod \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.041729 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-config-data\") pod \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\" (UID: \"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc\") " Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.050341 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-scripts" (OuterVolumeSpecName: "scripts") pod "4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" (UID: "4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.050531 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-kube-api-access-q77rd" (OuterVolumeSpecName: "kube-api-access-q77rd") pod "4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" (UID: "4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc"). InnerVolumeSpecName "kube-api-access-q77rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.074037 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66f99c579b-gh7sh" podUID="b78e2453-adc3-4139-97c5-9f8a977df1ee" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.095887 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-config-data" (OuterVolumeSpecName: "config-data") pod "4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" (UID: "4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.142883 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.142931 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q77rd\" (UniqueName: \"kubernetes.io/projected/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-kube-api-access-q77rd\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.142946 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.147115 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" (UID: "4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.244497 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.576884 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-22zjj" event={"ID":"4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc","Type":"ContainerDied","Data":"fe8aebc5e62fb36a8590bc5c4ad955205a3a78e0a3585c1dbccf4f064e236a4e"} Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.576988 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-22zjj" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.578027 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe8aebc5e62fb36a8590bc5c4ad955205a3a78e0a3585c1dbccf4f064e236a4e" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.721600 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 09:03:31 crc kubenswrapper[4573]: E1203 09:03:31.722055 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" containerName="nova-cell0-conductor-db-sync" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.722075 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" containerName="nova-cell0-conductor-db-sync" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.722277 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" containerName="nova-cell0-conductor-db-sync" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.722910 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.724938 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.737309 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-7b8bg" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.754330 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.856828 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a96a7087-9cec-427a-b088-0d900045bde3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.856969 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m8j2\" (UniqueName: \"kubernetes.io/projected/a96a7087-9cec-427a-b088-0d900045bde3-kube-api-access-5m8j2\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.857011 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a96a7087-9cec-427a-b088-0d900045bde3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.959407 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a96a7087-9cec-427a-b088-0d900045bde3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.959568 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m8j2\" (UniqueName: \"kubernetes.io/projected/a96a7087-9cec-427a-b088-0d900045bde3-kube-api-access-5m8j2\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.959600 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a96a7087-9cec-427a-b088-0d900045bde3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.965306 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a96a7087-9cec-427a-b088-0d900045bde3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.972193 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a96a7087-9cec-427a-b088-0d900045bde3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:31 crc kubenswrapper[4573]: I1203 09:03:31.983726 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m8j2\" (UniqueName: \"kubernetes.io/projected/a96a7087-9cec-427a-b088-0d900045bde3-kube-api-access-5m8j2\") pod \"nova-cell0-conductor-0\" (UID: \"a96a7087-9cec-427a-b088-0d900045bde3\") " pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:32 crc kubenswrapper[4573]: I1203 09:03:32.051531 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:32 crc kubenswrapper[4573]: I1203 09:03:32.629583 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.309866 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tmzcx"] Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.312111 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.326758 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tmzcx"] Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.394936 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-utilities\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.395030 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-catalog-content\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.395106 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4znmf\" (UniqueName: \"kubernetes.io/projected/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-kube-api-access-4znmf\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.496703 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-utilities\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.496828 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-catalog-content\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.496886 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4znmf\" (UniqueName: \"kubernetes.io/projected/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-kube-api-access-4znmf\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.497677 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-utilities\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.497837 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-catalog-content\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.517501 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4znmf\" (UniqueName: \"kubernetes.io/projected/f4d1c7fd-3939-4391-ae35-24bc7c2beaeb-kube-api-access-4znmf\") pod \"redhat-operators-tmzcx\" (UID: \"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb\") " pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.634862 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.658985 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a96a7087-9cec-427a-b088-0d900045bde3","Type":"ContainerStarted","Data":"f0e994e01fb05e0723266498d68941b714336d33ce222a1b782797d5c94ff4c8"} Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.659241 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a96a7087-9cec-427a-b088-0d900045bde3","Type":"ContainerStarted","Data":"7f75adfe92ea63e686881561bef758e3311806df9eef0a5e71c3acda3083f71f"} Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.660174 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.677221 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.677202932 podStartE2EDuration="2.677202932s" podCreationTimestamp="2025-12-03 09:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:03:33.675902646 +0000 UTC m=+1534.244281905" watchObservedRunningTime="2025-12-03 09:03:33.677202932 +0000 UTC m=+1534.245582191" Dec 03 09:03:33 crc kubenswrapper[4573]: I1203 09:03:33.979867 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tmzcx"] Dec 03 09:03:34 crc kubenswrapper[4573]: I1203 09:03:34.675369 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4d1c7fd-3939-4391-ae35-24bc7c2beaeb" containerID="e7096b8c7bb8f12a4065a7299648faed65c96b02076a1164634355ababb77d49" exitCode=0 Dec 03 09:03:34 crc kubenswrapper[4573]: I1203 09:03:34.675480 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmzcx" event={"ID":"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb","Type":"ContainerDied","Data":"e7096b8c7bb8f12a4065a7299648faed65c96b02076a1164634355ababb77d49"} Dec 03 09:03:34 crc kubenswrapper[4573]: I1203 09:03:34.675792 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmzcx" event={"ID":"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb","Type":"ContainerStarted","Data":"b8f01d58dd5901b3a6e154189d2e89a2187f0dc07dcb443a963755590675a06a"} Dec 03 09:03:37 crc kubenswrapper[4573]: I1203 09:03:37.095324 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 09:03:37 crc kubenswrapper[4573]: I1203 09:03:37.989605 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-vmz2j"] Dec 03 09:03:37 crc kubenswrapper[4573]: I1203 09:03:37.991418 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:37 crc kubenswrapper[4573]: I1203 09:03:37.996180 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 09:03:37 crc kubenswrapper[4573]: I1203 09:03:37.996525 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.003234 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vmz2j"] Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.127952 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.128167 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc77j\" (UniqueName: \"kubernetes.io/projected/f3e275e6-2937-4c08-8b15-537b25e00cb1-kube-api-access-fc77j\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.128287 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-config-data\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.128324 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-scripts\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.230938 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.231381 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc77j\" (UniqueName: \"kubernetes.io/projected/f3e275e6-2937-4c08-8b15-537b25e00cb1-kube-api-access-fc77j\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.231573 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-config-data\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.231699 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-scripts\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.250161 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.260945 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-config-data\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.285960 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-scripts\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.305063 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.308503 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.351413 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc77j\" (UniqueName: \"kubernetes.io/projected/f3e275e6-2937-4c08-8b15-537b25e00cb1-kube-api-access-fc77j\") pod \"nova-cell0-cell-mapping-vmz2j\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.352155 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.390488 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.441965 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-config-data\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.442136 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.442173 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hldm2\" (UniqueName: \"kubernetes.io/projected/01ed2237-985e-4599-b585-842558702f87-kube-api-access-hldm2\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.566185 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-config-data\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.567085 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.567248 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hldm2\" (UniqueName: \"kubernetes.io/projected/01ed2237-985e-4599-b585-842558702f87-kube-api-access-hldm2\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.569875 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.576039 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.599649 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.600129 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-config-data\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.600649 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.613588 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.636593 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.655061 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hldm2\" (UniqueName: \"kubernetes.io/projected/01ed2237-985e-4599-b585-842558702f87-kube-api-access-hldm2\") pod \"nova-scheduler-0\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.669898 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.669989 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q887\" (UniqueName: \"kubernetes.io/projected/142ca013-7220-454f-9e00-05af4f3f0ded-kube-api-access-2q887\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.670066 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.773320 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.773420 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q887\" (UniqueName: \"kubernetes.io/projected/142ca013-7220-454f-9e00-05af4f3f0ded-kube-api-access-2q887\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.773469 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.785736 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.788733 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.795787 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.839453 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q887\" (UniqueName: \"kubernetes.io/projected/142ca013-7220-454f-9e00-05af4f3f0ded-kube-api-access-2q887\") pod \"nova-cell1-novncproxy-0\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.906631 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.908496 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.916541 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.969206 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.970614 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.975193 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:03:38 crc kubenswrapper[4573]: I1203 09:03:38.991913 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.021283 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.060802 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.100837 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2c14a2a-018a-4190-8f7d-fc55004eaabb-logs\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.101630 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-config-data\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.101763 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hb2x\" (UniqueName: \"kubernetes.io/projected/b2c14a2a-018a-4190-8f7d-fc55004eaabb-kube-api-access-7hb2x\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.101905 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.102018 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-config-data\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.102177 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbc6h\" (UniqueName: \"kubernetes.io/projected/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-kube-api-access-lbc6h\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.102287 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-logs\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.102380 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.206187 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-config-data\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.206753 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hb2x\" (UniqueName: \"kubernetes.io/projected/b2c14a2a-018a-4190-8f7d-fc55004eaabb-kube-api-access-7hb2x\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.206799 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.206835 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-config-data\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.206857 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbc6h\" (UniqueName: \"kubernetes.io/projected/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-kube-api-access-lbc6h\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.206891 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-logs\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.206925 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.207035 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2c14a2a-018a-4190-8f7d-fc55004eaabb-logs\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.214814 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-logs\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.221448 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2c14a2a-018a-4190-8f7d-fc55004eaabb-logs\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.232285 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-config-data\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.251067 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-config-data\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.259159 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.267926 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.284067 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbc6h\" (UniqueName: \"kubernetes.io/projected/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-kube-api-access-lbc6h\") pod \"nova-metadata-0\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.293370 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hb2x\" (UniqueName: \"kubernetes.io/projected/b2c14a2a-018a-4190-8f7d-fc55004eaabb-kube-api-access-7hb2x\") pod \"nova-api-0\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.294146 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.377379 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.427700 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-lhxvs"] Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.481440 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.583131 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-lhxvs"] Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.651028 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.651115 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.651183 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4srnd\" (UniqueName: \"kubernetes.io/projected/8248677f-8e43-4fd1-92ed-560d826e6d60-kube-api-access-4srnd\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.651227 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.651354 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-config\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.651593 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-svc\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.756173 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-config\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.756453 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-svc\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.756547 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.756586 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.756643 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4srnd\" (UniqueName: \"kubernetes.io/projected/8248677f-8e43-4fd1-92ed-560d826e6d60-kube-api-access-4srnd\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.756678 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.765957 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-config\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.768789 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.768885 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-svc\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.770094 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.770137 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.797955 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4srnd\" (UniqueName: \"kubernetes.io/projected/8248677f-8e43-4fd1-92ed-560d826e6d60-kube-api-access-4srnd\") pod \"dnsmasq-dns-bccf8f775-lhxvs\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:39 crc kubenswrapper[4573]: I1203 09:03:39.907148 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.091879 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-vmz2j"] Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.129877 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.636270 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.648592 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.706593 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:03:40 crc kubenswrapper[4573]: W1203 09:03:40.724319 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8a62ff8_6759_4a8e_971c_c8ab8ce2561f.slice/crio-770e937ff002d24e811a73da8adf2ac3a68167f975729770cce81f648de4c4e5 WatchSource:0}: Error finding container 770e937ff002d24e811a73da8adf2ac3a68167f975729770cce81f648de4c4e5: Status 404 returned error can't find the container with id 770e937ff002d24e811a73da8adf2ac3a68167f975729770cce81f648de4c4e5 Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.823262 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f","Type":"ContainerStarted","Data":"770e937ff002d24e811a73da8adf2ac3a68167f975729770cce81f648de4c4e5"} Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.834919 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vmz2j" event={"ID":"f3e275e6-2937-4c08-8b15-537b25e00cb1","Type":"ContainerStarted","Data":"f050e023d9d1085a72c09420178f8232c10a3acb16222c9a30a70643ea37e1c4"} Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.834972 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vmz2j" event={"ID":"f3e275e6-2937-4c08-8b15-537b25e00cb1","Type":"ContainerStarted","Data":"3b5a993b055205ae55f30758888a0b60d286a64940eb87862313e7e481eb593c"} Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.850321 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"142ca013-7220-454f-9e00-05af4f3f0ded","Type":"ContainerStarted","Data":"758c82ba9a0c8f8bacbd96c0a18c906fb74398de713d977bfe8987a062235c0e"} Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.852374 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"01ed2237-985e-4599-b585-842558702f87","Type":"ContainerStarted","Data":"fcbd683d64ceb76b994a8cda2e39cc907891738a785863bb7d6e373b40badd67"} Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.853480 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b2c14a2a-018a-4190-8f7d-fc55004eaabb","Type":"ContainerStarted","Data":"0fde18d388c8fe5484f4d103ac4b27ac22f34b1708705af832fa0cfd8b846fa9"} Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.883841 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-vmz2j" podStartSLOduration=3.88381592 podStartE2EDuration="3.88381592s" podCreationTimestamp="2025-12-03 09:03:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:03:40.867737249 +0000 UTC m=+1541.436116508" watchObservedRunningTime="2025-12-03 09:03:40.88381592 +0000 UTC m=+1541.452195179" Dec 03 09:03:40 crc kubenswrapper[4573]: I1203 09:03:40.937124 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-lhxvs"] Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.146381 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s9sms"] Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.148163 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.154415 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.154620 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.185967 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s9sms"] Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.210977 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-config-data\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.211099 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxd56\" (UniqueName: \"kubernetes.io/projected/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-kube-api-access-lxd56\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.211175 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-scripts\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.237492 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.339788 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.340182 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-config-data\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.340335 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxd56\" (UniqueName: \"kubernetes.io/projected/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-kube-api-access-lxd56\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.340465 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-scripts\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.354038 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-config-data\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.354464 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.360090 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-scripts\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.368683 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxd56\" (UniqueName: \"kubernetes.io/projected/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-kube-api-access-lxd56\") pod \"nova-cell1-conductor-db-sync-s9sms\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:41 crc kubenswrapper[4573]: I1203 09:03:41.518529 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:03:42 crc kubenswrapper[4573]: I1203 09:03:42.093800 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" event={"ID":"8248677f-8e43-4fd1-92ed-560d826e6d60","Type":"ContainerStarted","Data":"18cdc7540bff79b8d406315a9abf428dea11eb97b58e3472946d4b079237bd3c"} Dec 03 09:03:42 crc kubenswrapper[4573]: I1203 09:03:42.094255 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" event={"ID":"8248677f-8e43-4fd1-92ed-560d826e6d60","Type":"ContainerStarted","Data":"732cd0a2e8caac2afcc63ff38227e64a34865c72d889adc88b088dc65d38455c"} Dec 03 09:03:42 crc kubenswrapper[4573]: I1203 09:03:42.384903 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s9sms"] Dec 03 09:03:43 crc kubenswrapper[4573]: I1203 09:03:43.074736 4573 generic.go:334] "Generic (PLEG): container finished" podID="8248677f-8e43-4fd1-92ed-560d826e6d60" containerID="18cdc7540bff79b8d406315a9abf428dea11eb97b58e3472946d4b079237bd3c" exitCode=0 Dec 03 09:03:43 crc kubenswrapper[4573]: I1203 09:03:43.078905 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" event={"ID":"8248677f-8e43-4fd1-92ed-560d826e6d60","Type":"ContainerDied","Data":"18cdc7540bff79b8d406315a9abf428dea11eb97b58e3472946d4b079237bd3c"} Dec 03 09:03:43 crc kubenswrapper[4573]: I1203 09:03:43.079329 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:43 crc kubenswrapper[4573]: I1203 09:03:43.079341 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" event={"ID":"8248677f-8e43-4fd1-92ed-560d826e6d60","Type":"ContainerStarted","Data":"4b7e20ae8779c2c6711539056f922aafacf8665fd530121e91f4181b96ed33b5"} Dec 03 09:03:43 crc kubenswrapper[4573]: I1203 09:03:43.104095 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s9sms" event={"ID":"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44","Type":"ContainerStarted","Data":"5b31ec267d0fa581174e02bec19006b16d2a3f12392424e35d0bad822f3acf27"} Dec 03 09:03:43 crc kubenswrapper[4573]: I1203 09:03:43.104181 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s9sms" event={"ID":"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44","Type":"ContainerStarted","Data":"b885eb16906ccef53b662d35f7ae182a44583953ab1cf05328e3c7f29698da42"} Dec 03 09:03:43 crc kubenswrapper[4573]: I1203 09:03:43.123419 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" podStartSLOduration=4.123396837 podStartE2EDuration="4.123396837s" podCreationTimestamp="2025-12-03 09:03:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:03:43.113181426 +0000 UTC m=+1543.681560685" watchObservedRunningTime="2025-12-03 09:03:43.123396837 +0000 UTC m=+1543.691776086" Dec 03 09:03:43 crc kubenswrapper[4573]: I1203 09:03:43.160753 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-s9sms" podStartSLOduration=2.160717543 podStartE2EDuration="2.160717543s" podCreationTimestamp="2025-12-03 09:03:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:03:43.141681619 +0000 UTC m=+1543.710060878" watchObservedRunningTime="2025-12-03 09:03:43.160717543 +0000 UTC m=+1543.729096802" Dec 03 09:03:44 crc kubenswrapper[4573]: I1203 09:03:44.113927 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:03:44 crc kubenswrapper[4573]: I1203 09:03:44.134522 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:03:44 crc kubenswrapper[4573]: I1203 09:03:44.922375 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 09:03:45 crc kubenswrapper[4573]: I1203 09:03:45.529846 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:03:45 crc kubenswrapper[4573]: I1203 09:03:45.931477 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:03:47 crc kubenswrapper[4573]: I1203 09:03:47.981644 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-66f99c579b-gh7sh" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.072661 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b8b5b6556-4zxxs"] Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.073004 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon-log" containerID="cri-o://39cb62ba4ad569a45049cde16bc1cf40005dce1875fe10337e5b5ae8bcb634a7" gracePeriod=30 Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.073564 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b8b5b6556-4zxxs" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" containerID="cri-o://d0c803805d4bd3808e5ee89977579fff41f3ca9da62cdfa598f77cba0d2c5e22" gracePeriod=30 Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.426336 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xrv4g"] Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.474398 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.483398 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xrv4g"] Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.626814 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-catalog-content\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.627166 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf45v\" (UniqueName: \"kubernetes.io/projected/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-kube-api-access-zf45v\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.628115 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-utilities\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.730974 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-catalog-content\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.731040 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf45v\" (UniqueName: \"kubernetes.io/projected/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-kube-api-access-zf45v\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.731188 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-utilities\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.731952 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-utilities\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.732094 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-catalog-content\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.777067 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf45v\" (UniqueName: \"kubernetes.io/projected/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-kube-api-access-zf45v\") pod \"certified-operators-xrv4g\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:48 crc kubenswrapper[4573]: I1203 09:03:48.872490 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:03:49 crc kubenswrapper[4573]: I1203 09:03:49.303116 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"142ca013-7220-454f-9e00-05af4f3f0ded","Type":"ContainerStarted","Data":"5babce96efa2d3c058b8fc06692f93399e8e3ae4be8fda57861915397e3d9bc6"} Dec 03 09:03:49 crc kubenswrapper[4573]: I1203 09:03:49.303495 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="142ca013-7220-454f-9e00-05af4f3f0ded" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://5babce96efa2d3c058b8fc06692f93399e8e3ae4be8fda57861915397e3d9bc6" gracePeriod=30 Dec 03 09:03:49 crc kubenswrapper[4573]: I1203 09:03:49.321347 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b2c14a2a-018a-4190-8f7d-fc55004eaabb","Type":"ContainerStarted","Data":"47588117147ca6a44198ed5f34f84e21d454adfd06e8b910348af22ec4065836"} Dec 03 09:03:49 crc kubenswrapper[4573]: I1203 09:03:49.376839 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.701860305 podStartE2EDuration="11.36602618s" podCreationTimestamp="2025-12-03 09:03:38 +0000 UTC" firstStartedPulling="2025-12-03 09:03:40.709659114 +0000 UTC m=+1541.278038383" lastFinishedPulling="2025-12-03 09:03:48.373824999 +0000 UTC m=+1548.942204258" observedRunningTime="2025-12-03 09:03:49.3598025 +0000 UTC m=+1549.928181759" watchObservedRunningTime="2025-12-03 09:03:49.36602618 +0000 UTC m=+1549.934405439" Dec 03 09:03:49 crc kubenswrapper[4573]: I1203 09:03:49.652026 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xrv4g"] Dec 03 09:03:49 crc kubenswrapper[4573]: W1203 09:03:49.698727 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf48f1b5a_0f0f_4491_bab1_6b602cfefb03.slice/crio-084daedbd5c9ce61014f6776c985169279eb4bc5f18dab29823b9875e42cc4a8 WatchSource:0}: Error finding container 084daedbd5c9ce61014f6776c985169279eb4bc5f18dab29823b9875e42cc4a8: Status 404 returned error can't find the container with id 084daedbd5c9ce61014f6776c985169279eb4bc5f18dab29823b9875e42cc4a8 Dec 03 09:03:49 crc kubenswrapper[4573]: I1203 09:03:49.912037 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.055348 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-79ggx"] Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.055778 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" podUID="b9a834f3-5601-48b5-982a-0ab476605db5" containerName="dnsmasq-dns" containerID="cri-o://bb16388dcadba35686654bf424cad2e59956e899e5746a662a72cd0a6d693540" gracePeriod=10 Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.345871 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f","Type":"ContainerStarted","Data":"146948da1064794d1e2d631e684c6054c37a8a0d99dd483594c3da25031fc222"} Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.357395 4573 generic.go:334] "Generic (PLEG): container finished" podID="b9a834f3-5601-48b5-982a-0ab476605db5" containerID="bb16388dcadba35686654bf424cad2e59956e899e5746a662a72cd0a6d693540" exitCode=0 Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.357488 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" event={"ID":"b9a834f3-5601-48b5-982a-0ab476605db5","Type":"ContainerDied","Data":"bb16388dcadba35686654bf424cad2e59956e899e5746a662a72cd0a6d693540"} Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.372895 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"01ed2237-985e-4599-b585-842558702f87","Type":"ContainerStarted","Data":"5db22f7d546abe1fa78ac7a4427f78a6d3425c8edb1499b37ee171da9e8b10f1"} Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.393476 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrv4g" event={"ID":"f48f1b5a-0f0f-4491-bab1-6b602cfefb03","Type":"ContainerStarted","Data":"084daedbd5c9ce61014f6776c985169279eb4bc5f18dab29823b9875e42cc4a8"} Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.417934 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.211531805 podStartE2EDuration="12.417908273s" podCreationTimestamp="2025-12-03 09:03:38 +0000 UTC" firstStartedPulling="2025-12-03 09:03:40.167441591 +0000 UTC m=+1540.735820850" lastFinishedPulling="2025-12-03 09:03:48.373818059 +0000 UTC m=+1548.942197318" observedRunningTime="2025-12-03 09:03:50.402414516 +0000 UTC m=+1550.970793775" watchObservedRunningTime="2025-12-03 09:03:50.417908273 +0000 UTC m=+1550.986287532" Dec 03 09:03:50 crc kubenswrapper[4573]: I1203 09:03:50.999294 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.034828 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-config\") pod \"b9a834f3-5601-48b5-982a-0ab476605db5\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.034920 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-swift-storage-0\") pod \"b9a834f3-5601-48b5-982a-0ab476605db5\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.035083 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-nb\") pod \"b9a834f3-5601-48b5-982a-0ab476605db5\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.035130 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-sb\") pod \"b9a834f3-5601-48b5-982a-0ab476605db5\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.035212 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-svc\") pod \"b9a834f3-5601-48b5-982a-0ab476605db5\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.035258 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hnmg\" (UniqueName: \"kubernetes.io/projected/b9a834f3-5601-48b5-982a-0ab476605db5-kube-api-access-4hnmg\") pod \"b9a834f3-5601-48b5-982a-0ab476605db5\" (UID: \"b9a834f3-5601-48b5-982a-0ab476605db5\") " Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.049952 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a834f3-5601-48b5-982a-0ab476605db5-kube-api-access-4hnmg" (OuterVolumeSpecName: "kube-api-access-4hnmg") pod "b9a834f3-5601-48b5-982a-0ab476605db5" (UID: "b9a834f3-5601-48b5-982a-0ab476605db5"). InnerVolumeSpecName "kube-api-access-4hnmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.205132 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hnmg\" (UniqueName: \"kubernetes.io/projected/b9a834f3-5601-48b5-982a-0ab476605db5-kube-api-access-4hnmg\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.237376 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b9a834f3-5601-48b5-982a-0ab476605db5" (UID: "b9a834f3-5601-48b5-982a-0ab476605db5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.240359 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-config" (OuterVolumeSpecName: "config") pod "b9a834f3-5601-48b5-982a-0ab476605db5" (UID: "b9a834f3-5601-48b5-982a-0ab476605db5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.312763 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.312804 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.320191 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b9a834f3-5601-48b5-982a-0ab476605db5" (UID: "b9a834f3-5601-48b5-982a-0ab476605db5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.351710 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b9a834f3-5601-48b5-982a-0ab476605db5" (UID: "b9a834f3-5601-48b5-982a-0ab476605db5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.377069 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b9a834f3-5601-48b5-982a-0ab476605db5" (UID: "b9a834f3-5601-48b5-982a-0ab476605db5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.420654 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.420683 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.420694 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9a834f3-5601-48b5-982a-0ab476605db5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.432307 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f","Type":"ContainerStarted","Data":"b1d3f69b7328466b4369c0a23be8aea1bc25ee5bb93369055b0f7fef85bd10a5"} Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.432521 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerName="nova-metadata-log" containerID="cri-o://146948da1064794d1e2d631e684c6054c37a8a0d99dd483594c3da25031fc222" gracePeriod=30 Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.433043 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerName="nova-metadata-metadata" containerID="cri-o://b1d3f69b7328466b4369c0a23be8aea1bc25ee5bb93369055b0f7fef85bd10a5" gracePeriod=30 Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.438634 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" event={"ID":"b9a834f3-5601-48b5-982a-0ab476605db5","Type":"ContainerDied","Data":"f3800725ea7fe587487ff6f8a06d2381b616d0a05f1f555811bbab9da2653d8b"} Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.438701 4573 scope.go:117] "RemoveContainer" containerID="bb16388dcadba35686654bf424cad2e59956e899e5746a662a72cd0a6d693540" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.438844 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-79ggx" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.447797 4573 generic.go:334] "Generic (PLEG): container finished" podID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerID="83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53" exitCode=0 Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.447878 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrv4g" event={"ID":"f48f1b5a-0f0f-4491-bab1-6b602cfefb03","Type":"ContainerDied","Data":"83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53"} Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.464855 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=5.81696168 podStartE2EDuration="13.464837478s" podCreationTimestamp="2025-12-03 09:03:38 +0000 UTC" firstStartedPulling="2025-12-03 09:03:40.728222054 +0000 UTC m=+1541.296601313" lastFinishedPulling="2025-12-03 09:03:48.376097862 +0000 UTC m=+1548.944477111" observedRunningTime="2025-12-03 09:03:51.463378398 +0000 UTC m=+1552.031757657" watchObservedRunningTime="2025-12-03 09:03:51.464837478 +0000 UTC m=+1552.033216737" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.465146 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b2c14a2a-018a-4190-8f7d-fc55004eaabb","Type":"ContainerStarted","Data":"b986037d887208c104d14de970e22d137e4a6987234e42c13309978841fe05cd"} Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.503121 4573 scope.go:117] "RemoveContainer" containerID="82d3e11445e4fcc3789cf87aedc48577465aa73fc4f5f6e9b073d2282c3bab87" Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.536496 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-79ggx"] Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.550935 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-79ggx"] Dec 03 09:03:51 crc kubenswrapper[4573]: I1203 09:03:51.568452 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=5.871284633 podStartE2EDuration="13.568428545s" podCreationTimestamp="2025-12-03 09:03:38 +0000 UTC" firstStartedPulling="2025-12-03 09:03:40.678858487 +0000 UTC m=+1541.247237746" lastFinishedPulling="2025-12-03 09:03:48.376002399 +0000 UTC m=+1548.944381658" observedRunningTime="2025-12-03 09:03:51.547792638 +0000 UTC m=+1552.116171887" watchObservedRunningTime="2025-12-03 09:03:51.568428545 +0000 UTC m=+1552.136807804" Dec 03 09:03:52 crc kubenswrapper[4573]: I1203 09:03:52.053713 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a834f3-5601-48b5-982a-0ab476605db5" path="/var/lib/kubelet/pods/b9a834f3-5601-48b5-982a-0ab476605db5/volumes" Dec 03 09:03:52 crc kubenswrapper[4573]: I1203 09:03:52.498951 4573 generic.go:334] "Generic (PLEG): container finished" podID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerID="146948da1064794d1e2d631e684c6054c37a8a0d99dd483594c3da25031fc222" exitCode=143 Dec 03 09:03:52 crc kubenswrapper[4573]: I1203 09:03:52.499035 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f","Type":"ContainerDied","Data":"146948da1064794d1e2d631e684c6054c37a8a0d99dd483594c3da25031fc222"} Dec 03 09:03:52 crc kubenswrapper[4573]: I1203 09:03:52.513372 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerID="d0c803805d4bd3808e5ee89977579fff41f3ca9da62cdfa598f77cba0d2c5e22" exitCode=0 Dec 03 09:03:52 crc kubenswrapper[4573]: I1203 09:03:52.514170 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerDied","Data":"d0c803805d4bd3808e5ee89977579fff41f3ca9da62cdfa598f77cba0d2c5e22"} Dec 03 09:03:52 crc kubenswrapper[4573]: I1203 09:03:52.514248 4573 scope.go:117] "RemoveContainer" containerID="b8aa7ba7c95bd0a8692253b9b93fecbaa60bed28d93ea2c5355468e1a8ef3047" Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.212412 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.212631 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="811326ff-3ff6-4d68-9c4f-4c356f09d6b2" containerName="kube-state-metrics" containerID="cri-o://64ce8654f051565f7b36c0dede7b298fbebd6eb40cf6baef743ad2b103898f4a" gracePeriod=30 Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.534380 4573 generic.go:334] "Generic (PLEG): container finished" podID="811326ff-3ff6-4d68-9c4f-4c356f09d6b2" containerID="64ce8654f051565f7b36c0dede7b298fbebd6eb40cf6baef743ad2b103898f4a" exitCode=2 Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.534462 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"811326ff-3ff6-4d68-9c4f-4c356f09d6b2","Type":"ContainerDied","Data":"64ce8654f051565f7b36c0dede7b298fbebd6eb40cf6baef743ad2b103898f4a"} Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.536643 4573 generic.go:334] "Generic (PLEG): container finished" podID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerID="b1d3f69b7328466b4369c0a23be8aea1bc25ee5bb93369055b0f7fef85bd10a5" exitCode=0 Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.536682 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f","Type":"ContainerDied","Data":"b1d3f69b7328466b4369c0a23be8aea1bc25ee5bb93369055b0f7fef85bd10a5"} Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.540846 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrv4g" event={"ID":"f48f1b5a-0f0f-4491-bab1-6b602cfefb03","Type":"ContainerStarted","Data":"076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af"} Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.786512 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.795684 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="811326ff-3ff6-4d68-9c4f-4c356f09d6b2" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": dial tcp 10.217.0.103:8081: connect: connection refused" Dec 03 09:03:53 crc kubenswrapper[4573]: I1203 09:03:53.971343 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:03:54 crc kubenswrapper[4573]: I1203 09:03:54.295083 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 09:03:54 crc kubenswrapper[4573]: I1203 09:03:54.295568 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 09:03:54 crc kubenswrapper[4573]: I1203 09:03:54.560504 4573 generic.go:334] "Generic (PLEG): container finished" podID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerID="076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af" exitCode=0 Dec 03 09:03:54 crc kubenswrapper[4573]: I1203 09:03:54.560553 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrv4g" event={"ID":"f48f1b5a-0f0f-4491-bab1-6b602cfefb03","Type":"ContainerDied","Data":"076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af"} Dec 03 09:03:55 crc kubenswrapper[4573]: I1203 09:03:55.573226 4573 generic.go:334] "Generic (PLEG): container finished" podID="f3e275e6-2937-4c08-8b15-537b25e00cb1" containerID="f050e023d9d1085a72c09420178f8232c10a3acb16222c9a30a70643ea37e1c4" exitCode=0 Dec 03 09:03:55 crc kubenswrapper[4573]: I1203 09:03:55.573317 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vmz2j" event={"ID":"f3e275e6-2937-4c08-8b15-537b25e00cb1","Type":"ContainerDied","Data":"f050e023d9d1085a72c09420178f8232c10a3acb16222c9a30a70643ea37e1c4"} Dec 03 09:03:55 crc kubenswrapper[4573]: I1203 09:03:55.892020 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:03:55 crc kubenswrapper[4573]: I1203 09:03:55.892353 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="ceilometer-central-agent" containerID="cri-o://b5ebe341f35a6457c3af554054a07d18d857a6e65d782fe67d07eea441859d27" gracePeriod=30 Dec 03 09:03:55 crc kubenswrapper[4573]: I1203 09:03:55.892470 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="proxy-httpd" containerID="cri-o://be7b5ba9dcb78e51473052228d392d38368303937453aaf4de98c0c173805179" gracePeriod=30 Dec 03 09:03:55 crc kubenswrapper[4573]: I1203 09:03:55.892507 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="sg-core" containerID="cri-o://0fee465fc39dc7de24f61324f3f6b461424cc50d89e57c0b496e59334bd9b664" gracePeriod=30 Dec 03 09:03:55 crc kubenswrapper[4573]: I1203 09:03:55.892536 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="ceilometer-notification-agent" containerID="cri-o://25248762cbd521ef8ef4803fd45cc8694960f9640ec665a520f524a5b23ada9c" gracePeriod=30 Dec 03 09:03:56 crc kubenswrapper[4573]: I1203 09:03:56.615206 4573 generic.go:334] "Generic (PLEG): container finished" podID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerID="be7b5ba9dcb78e51473052228d392d38368303937453aaf4de98c0c173805179" exitCode=0 Dec 03 09:03:56 crc kubenswrapper[4573]: I1203 09:03:56.615442 4573 generic.go:334] "Generic (PLEG): container finished" podID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerID="0fee465fc39dc7de24f61324f3f6b461424cc50d89e57c0b496e59334bd9b664" exitCode=2 Dec 03 09:03:56 crc kubenswrapper[4573]: I1203 09:03:56.615454 4573 generic.go:334] "Generic (PLEG): container finished" podID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerID="b5ebe341f35a6457c3af554054a07d18d857a6e65d782fe67d07eea441859d27" exitCode=0 Dec 03 09:03:56 crc kubenswrapper[4573]: I1203 09:03:56.616471 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerDied","Data":"be7b5ba9dcb78e51473052228d392d38368303937453aaf4de98c0c173805179"} Dec 03 09:03:56 crc kubenswrapper[4573]: I1203 09:03:56.616543 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerDied","Data":"0fee465fc39dc7de24f61324f3f6b461424cc50d89e57c0b496e59334bd9b664"} Dec 03 09:03:56 crc kubenswrapper[4573]: I1203 09:03:56.616574 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerDied","Data":"b5ebe341f35a6457c3af554054a07d18d857a6e65d782fe67d07eea441859d27"} Dec 03 09:03:58 crc kubenswrapper[4573]: I1203 09:03:58.635976 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s9sms" event={"ID":"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44","Type":"ContainerDied","Data":"5b31ec267d0fa581174e02bec19006b16d2a3f12392424e35d0bad822f3acf27"} Dec 03 09:03:58 crc kubenswrapper[4573]: I1203 09:03:58.636918 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" containerID="5b31ec267d0fa581174e02bec19006b16d2a3f12392424e35d0bad822f3acf27" exitCode=0 Dec 03 09:03:58 crc kubenswrapper[4573]: I1203 09:03:58.786686 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 09:03:58 crc kubenswrapper[4573]: I1203 09:03:58.821275 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 09:03:59 crc kubenswrapper[4573]: I1203 09:03:59.379191 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 09:03:59 crc kubenswrapper[4573]: I1203 09:03:59.379236 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 09:03:59 crc kubenswrapper[4573]: I1203 09:03:59.678648 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 09:04:00 crc kubenswrapper[4573]: I1203 09:04:00.489409 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:04:00 crc kubenswrapper[4573]: I1203 09:04:00.491022 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:04:02 crc kubenswrapper[4573]: I1203 09:04:02.713397 4573 generic.go:334] "Generic (PLEG): container finished" podID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerID="25248762cbd521ef8ef4803fd45cc8694960f9640ec665a520f524a5b23ada9c" exitCode=0 Dec 03 09:04:02 crc kubenswrapper[4573]: I1203 09:04:02.713452 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerDied","Data":"25248762cbd521ef8ef4803fd45cc8694960f9640ec665a520f524a5b23ada9c"} Dec 03 09:04:02 crc kubenswrapper[4573]: E1203 09:04:02.976323 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 09:04:02 crc kubenswrapper[4573]: E1203 09:04:02.976840 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4znmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-tmzcx_openshift-marketplace(f4d1c7fd-3939-4391-ae35-24bc7c2beaeb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 09:04:02 crc kubenswrapper[4573]: E1203 09:04:02.978704 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-tmzcx" podUID="f4d1c7fd-3939-4391-ae35-24bc7c2beaeb" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.123711 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.130114 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.257771 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-scripts\") pod \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.257850 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-scripts\") pod \"f3e275e6-2937-4c08-8b15-537b25e00cb1\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.257941 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc77j\" (UniqueName: \"kubernetes.io/projected/f3e275e6-2937-4c08-8b15-537b25e00cb1-kube-api-access-fc77j\") pod \"f3e275e6-2937-4c08-8b15-537b25e00cb1\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.257963 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-combined-ca-bundle\") pod \"f3e275e6-2937-4c08-8b15-537b25e00cb1\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.258013 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-config-data\") pod \"f3e275e6-2937-4c08-8b15-537b25e00cb1\" (UID: \"f3e275e6-2937-4c08-8b15-537b25e00cb1\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.258071 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-config-data\") pod \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.258132 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxd56\" (UniqueName: \"kubernetes.io/projected/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-kube-api-access-lxd56\") pod \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.258190 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-combined-ca-bundle\") pod \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\" (UID: \"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.271368 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-scripts" (OuterVolumeSpecName: "scripts") pod "f3e275e6-2937-4c08-8b15-537b25e00cb1" (UID: "f3e275e6-2937-4c08-8b15-537b25e00cb1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.274955 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-kube-api-access-lxd56" (OuterVolumeSpecName: "kube-api-access-lxd56") pod "f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" (UID: "f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44"). InnerVolumeSpecName "kube-api-access-lxd56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.282437 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e275e6-2937-4c08-8b15-537b25e00cb1-kube-api-access-fc77j" (OuterVolumeSpecName: "kube-api-access-fc77j") pod "f3e275e6-2937-4c08-8b15-537b25e00cb1" (UID: "f3e275e6-2937-4c08-8b15-537b25e00cb1"). InnerVolumeSpecName "kube-api-access-fc77j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.285745 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-scripts" (OuterVolumeSpecName: "scripts") pod "f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" (UID: "f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.309424 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" (UID: "f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.317492 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3e275e6-2937-4c08-8b15-537b25e00cb1" (UID: "f3e275e6-2937-4c08-8b15-537b25e00cb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.325284 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-config-data" (OuterVolumeSpecName: "config-data") pod "f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" (UID: "f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.340486 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-config-data" (OuterVolumeSpecName: "config-data") pod "f3e275e6-2937-4c08-8b15-537b25e00cb1" (UID: "f3e275e6-2937-4c08-8b15-537b25e00cb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.360853 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.360885 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.360895 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc77j\" (UniqueName: \"kubernetes.io/projected/f3e275e6-2937-4c08-8b15-537b25e00cb1-kube-api-access-fc77j\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.360910 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.360921 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3e275e6-2937-4c08-8b15-537b25e00cb1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.360930 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.360938 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxd56\" (UniqueName: \"kubernetes.io/projected/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-kube-api-access-lxd56\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.360947 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.462859 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.549764 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.572081 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcfpr\" (UniqueName: \"kubernetes.io/projected/811326ff-3ff6-4d68-9c4f-4c356f09d6b2-kube-api-access-qcfpr\") pod \"811326ff-3ff6-4d68-9c4f-4c356f09d6b2\" (UID: \"811326ff-3ff6-4d68-9c4f-4c356f09d6b2\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.580665 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/811326ff-3ff6-4d68-9c4f-4c356f09d6b2-kube-api-access-qcfpr" (OuterVolumeSpecName: "kube-api-access-qcfpr") pod "811326ff-3ff6-4d68-9c4f-4c356f09d6b2" (UID: "811326ff-3ff6-4d68-9c4f-4c356f09d6b2"). InnerVolumeSpecName "kube-api-access-qcfpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.673402 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-config-data\") pod \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.673583 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-logs\") pod \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.673608 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-combined-ca-bundle\") pod \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.673745 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbc6h\" (UniqueName: \"kubernetes.io/projected/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-kube-api-access-lbc6h\") pod \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\" (UID: \"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.674227 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcfpr\" (UniqueName: \"kubernetes.io/projected/811326ff-3ff6-4d68-9c4f-4c356f09d6b2-kube-api-access-qcfpr\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.674610 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-logs" (OuterVolumeSpecName: "logs") pod "f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" (UID: "f8a62ff8-6759-4a8e-971c-c8ab8ce2561f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.693763 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-kube-api-access-lbc6h" (OuterVolumeSpecName: "kube-api-access-lbc6h") pod "f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" (UID: "f8a62ff8-6759-4a8e-971c-c8ab8ce2561f"). InnerVolumeSpecName "kube-api-access-lbc6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.735087 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.735250 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s9sms" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.736238 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s9sms" event={"ID":"f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44","Type":"ContainerDied","Data":"b885eb16906ccef53b662d35f7ae182a44583953ab1cf05328e3c7f29698da42"} Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.736278 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b885eb16906ccef53b662d35f7ae182a44583953ab1cf05328e3c7f29698da42" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.744751 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"441a63f9-caa0-4f25-8def-194be8fd64a6","Type":"ContainerDied","Data":"faf7b08b1d15f46454874a7248255627e413768aaa00da2e111f106fb98308a5"} Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.745085 4573 scope.go:117] "RemoveContainer" containerID="be7b5ba9dcb78e51473052228d392d38368303937453aaf4de98c0c173805179" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.754852 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"811326ff-3ff6-4d68-9c4f-4c356f09d6b2","Type":"ContainerDied","Data":"4526b531e75fefcfc7a00490340faba2f23a2806e9df8ac46c1a37937b6c02e4"} Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.754942 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.758842 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8a62ff8-6759-4a8e-971c-c8ab8ce2561f","Type":"ContainerDied","Data":"770e937ff002d24e811a73da8adf2ac3a68167f975729770cce81f648de4c4e5"} Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.758931 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.766317 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-vmz2j" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.769904 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" (UID: "f8a62ff8-6759-4a8e-971c-c8ab8ce2561f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.770639 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-vmz2j" event={"ID":"f3e275e6-2937-4c08-8b15-537b25e00cb1","Type":"ContainerDied","Data":"3b5a993b055205ae55f30758888a0b60d286a64940eb87862313e7e481eb593c"} Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.775061 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b5a993b055205ae55f30758888a0b60d286a64940eb87862313e7e481eb593c" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.777022 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbc6h\" (UniqueName: \"kubernetes.io/projected/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-kube-api-access-lbc6h\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.777078 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.777090 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: E1203 09:04:03.800264 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-tmzcx" podUID="f4d1c7fd-3939-4391-ae35-24bc7c2beaeb" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.800554 4573 scope.go:117] "RemoveContainer" containerID="0fee465fc39dc7de24f61324f3f6b461424cc50d89e57c0b496e59334bd9b664" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.813912 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-config-data" (OuterVolumeSpecName: "config-data") pod "f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" (UID: "f8a62ff8-6759-4a8e-971c-c8ab8ce2561f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.881635 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-sg-core-conf-yaml\") pod \"441a63f9-caa0-4f25-8def-194be8fd64a6\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.888122 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-run-httpd\") pod \"441a63f9-caa0-4f25-8def-194be8fd64a6\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.888413 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-config-data\") pod \"441a63f9-caa0-4f25-8def-194be8fd64a6\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.888557 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-combined-ca-bundle\") pod \"441a63f9-caa0-4f25-8def-194be8fd64a6\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.888772 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-scripts\") pod \"441a63f9-caa0-4f25-8def-194be8fd64a6\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.889623 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-log-httpd\") pod \"441a63f9-caa0-4f25-8def-194be8fd64a6\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.889871 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw6f9\" (UniqueName: \"kubernetes.io/projected/441a63f9-caa0-4f25-8def-194be8fd64a6-kube-api-access-jw6f9\") pod \"441a63f9-caa0-4f25-8def-194be8fd64a6\" (UID: \"441a63f9-caa0-4f25-8def-194be8fd64a6\") " Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.892113 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "441a63f9-caa0-4f25-8def-194be8fd64a6" (UID: "441a63f9-caa0-4f25-8def-194be8fd64a6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.895709 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "441a63f9-caa0-4f25-8def-194be8fd64a6" (UID: "441a63f9-caa0-4f25-8def-194be8fd64a6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.928541 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.928685 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.928747 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/441a63f9-caa0-4f25-8def-194be8fd64a6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.942093 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/441a63f9-caa0-4f25-8def-194be8fd64a6-kube-api-access-jw6f9" (OuterVolumeSpecName: "kube-api-access-jw6f9") pod "441a63f9-caa0-4f25-8def-194be8fd64a6" (UID: "441a63f9-caa0-4f25-8def-194be8fd64a6"). InnerVolumeSpecName "kube-api-access-jw6f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.950239 4573 scope.go:117] "RemoveContainer" containerID="25248762cbd521ef8ef4803fd45cc8694960f9640ec665a520f524a5b23ada9c" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.951036 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-scripts" (OuterVolumeSpecName: "scripts") pod "441a63f9-caa0-4f25-8def-194be8fd64a6" (UID: "441a63f9-caa0-4f25-8def-194be8fd64a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.951109 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "441a63f9-caa0-4f25-8def-194be8fd64a6" (UID: "441a63f9-caa0-4f25-8def-194be8fd64a6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:03 crc kubenswrapper[4573]: I1203 09:04:03.976031 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.034057 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw6f9\" (UniqueName: \"kubernetes.io/projected/441a63f9-caa0-4f25-8def-194be8fd64a6-kube-api-access-jw6f9\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.034359 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.034437 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.069255 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "441a63f9-caa0-4f25-8def-194be8fd64a6" (UID: "441a63f9-caa0-4f25-8def-194be8fd64a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.135928 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.165993 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-config-data" (OuterVolumeSpecName: "config-data") pod "441a63f9-caa0-4f25-8def-194be8fd64a6" (UID: "441a63f9-caa0-4f25-8def-194be8fd64a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227158 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227211 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227727 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="ceilometer-notification-agent" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227748 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="ceilometer-notification-agent" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227760 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerName="nova-metadata-metadata" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227767 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerName="nova-metadata-metadata" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227779 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="sg-core" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227787 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="sg-core" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227798 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e275e6-2937-4c08-8b15-537b25e00cb1" containerName="nova-manage" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227805 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e275e6-2937-4c08-8b15-537b25e00cb1" containerName="nova-manage" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227816 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a834f3-5601-48b5-982a-0ab476605db5" containerName="dnsmasq-dns" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227822 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a834f3-5601-48b5-982a-0ab476605db5" containerName="dnsmasq-dns" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227836 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="ceilometer-central-agent" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227843 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="ceilometer-central-agent" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227852 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="proxy-httpd" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227858 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="proxy-httpd" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227869 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerName="nova-metadata-log" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227875 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerName="nova-metadata-log" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227889 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a834f3-5601-48b5-982a-0ab476605db5" containerName="init" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227895 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a834f3-5601-48b5-982a-0ab476605db5" containerName="init" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227906 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" containerName="nova-cell1-conductor-db-sync" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227912 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" containerName="nova-cell1-conductor-db-sync" Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.227928 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="811326ff-3ff6-4d68-9c4f-4c356f09d6b2" containerName="kube-state-metrics" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.227935 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="811326ff-3ff6-4d68-9c4f-4c356f09d6b2" containerName="kube-state-metrics" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228143 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="ceilometer-notification-agent" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228163 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" containerName="nova-cell1-conductor-db-sync" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228172 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3e275e6-2937-4c08-8b15-537b25e00cb1" containerName="nova-manage" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228181 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="ceilometer-central-agent" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228193 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerName="nova-metadata-log" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228203 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="811326ff-3ff6-4d68-9c4f-4c356f09d6b2" containerName="kube-state-metrics" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228212 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" containerName="nova-metadata-metadata" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228225 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="sg-core" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228234 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a834f3-5601-48b5-982a-0ab476605db5" containerName="dnsmasq-dns" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228244 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" containerName="proxy-httpd" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.228922 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.229011 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.243765 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/441a63f9-caa0-4f25-8def-194be8fd64a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.255472 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.255720 4573 scope.go:117] "RemoveContainer" containerID="b5ebe341f35a6457c3af554054a07d18d857a6e65d782fe67d07eea441859d27" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.255915 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.280204 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.292333 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.301520 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.303488 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.306603 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.306828 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.318795 4573 scope.go:117] "RemoveContainer" containerID="64ce8654f051565f7b36c0dede7b298fbebd6eb40cf6baef743ad2b103898f4a" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.345864 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.345935 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.345987 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.346105 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67wbz\" (UniqueName: \"kubernetes.io/projected/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-api-access-67wbz\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.363274 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.373877 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.375287 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.379626 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.380580 4573 scope.go:117] "RemoveContainer" containerID="b1d3f69b7328466b4369c0a23be8aea1bc25ee5bb93369055b0f7fef85bd10a5" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.411124 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.427015 4573 scope.go:117] "RemoveContainer" containerID="146948da1064794d1e2d631e684c6054c37a8a0d99dd483594c3da25031fc222" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.447454 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.447777 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2g6s\" (UniqueName: \"kubernetes.io/projected/2227cd84-de1e-4b88-8712-dbd6471e6c67-kube-api-access-k2g6s\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.447861 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.447989 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f1297d-f47b-4ac3-a9ba-3466972af97e-logs\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.448098 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2227cd84-de1e-4b88-8712-dbd6471e6c67-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.448201 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr6cw\" (UniqueName: \"kubernetes.io/projected/d6f1297d-f47b-4ac3-a9ba-3466972af97e-kube-api-access-sr6cw\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.448378 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67wbz\" (UniqueName: \"kubernetes.io/projected/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-api-access-67wbz\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.448540 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.448649 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-config-data\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.448893 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2227cd84-de1e-4b88-8712-dbd6471e6c67-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.448992 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.449210 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.447675 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.449711 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-log" containerID="cri-o://47588117147ca6a44198ed5f34f84e21d454adfd06e8b910348af22ec4065836" gracePeriod=30 Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.452555 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-api" containerID="cri-o://b986037d887208c104d14de970e22d137e4a6987234e42c13309978841fe05cd" gracePeriod=30 Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.463770 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.472133 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.472664 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.495864 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67wbz\" (UniqueName: \"kubernetes.io/projected/e50ae89c-c3c5-44a5-9222-b14ddbadc294-kube-api-access-67wbz\") pod \"kube-state-metrics-0\" (UID: \"e50ae89c-c3c5-44a5-9222-b14ddbadc294\") " pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.498206 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.498426 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="01ed2237-985e-4599-b585-842558702f87" containerName="nova-scheduler-scheduler" containerID="cri-o://5db22f7d546abe1fa78ac7a4427f78a6d3425c8edb1499b37ee171da9e8b10f1" gracePeriod=30 Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.541797 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: E1203 09:04:04.543036 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-sr6cw logs nova-metadata-tls-certs], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/nova-metadata-0" podUID="d6f1297d-f47b-4ac3-a9ba-3466972af97e" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.554132 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-config-data\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.554185 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2227cd84-de1e-4b88-8712-dbd6471e6c67-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.554220 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.554272 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.554290 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2g6s\" (UniqueName: \"kubernetes.io/projected/2227cd84-de1e-4b88-8712-dbd6471e6c67-kube-api-access-k2g6s\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.554313 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f1297d-f47b-4ac3-a9ba-3466972af97e-logs\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.554336 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2227cd84-de1e-4b88-8712-dbd6471e6c67-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.554358 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr6cw\" (UniqueName: \"kubernetes.io/projected/d6f1297d-f47b-4ac3-a9ba-3466972af97e-kube-api-access-sr6cw\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.555621 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f1297d-f47b-4ac3-a9ba-3466972af97e-logs\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.560674 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.560709 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-config-data\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.561138 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2227cd84-de1e-4b88-8712-dbd6471e6c67-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.562945 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.568628 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2227cd84-de1e-4b88-8712-dbd6471e6c67-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.570410 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.575678 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr6cw\" (UniqueName: \"kubernetes.io/projected/d6f1297d-f47b-4ac3-a9ba-3466972af97e-kube-api-access-sr6cw\") pod \"nova-metadata-0\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.581571 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2g6s\" (UniqueName: \"kubernetes.io/projected/2227cd84-de1e-4b88-8712-dbd6471e6c67-kube-api-access-k2g6s\") pod \"nova-cell1-conductor-0\" (UID: \"2227cd84-de1e-4b88-8712-dbd6471e6c67\") " pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.722581 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.824767 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.851725 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrv4g" event={"ID":"f48f1b5a-0f0f-4491-bab1-6b602cfefb03","Type":"ContainerStarted","Data":"b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4"} Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.887445 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.926187 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.930874 4573 generic.go:334] "Generic (PLEG): container finished" podID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerID="47588117147ca6a44198ed5f34f84e21d454adfd06e8b910348af22ec4065836" exitCode=143 Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.930975 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b2c14a2a-018a-4190-8f7d-fc55004eaabb","Type":"ContainerDied","Data":"47588117147ca6a44198ed5f34f84e21d454adfd06e8b910348af22ec4065836"} Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.937603 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.947208 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xrv4g" podStartSLOduration=4.937725722 podStartE2EDuration="16.94718298s" podCreationTimestamp="2025-12-03 09:03:48 +0000 UTC" firstStartedPulling="2025-12-03 09:03:51.451086541 +0000 UTC m=+1552.019465800" lastFinishedPulling="2025-12-03 09:04:03.460543809 +0000 UTC m=+1564.028923058" observedRunningTime="2025-12-03 09:04:04.911115108 +0000 UTC m=+1565.479494367" watchObservedRunningTime="2025-12-03 09:04:04.94718298 +0000 UTC m=+1565.515562239" Dec 03 09:04:04 crc kubenswrapper[4573]: I1203 09:04:04.970663 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.016543 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.016732 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.021768 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.023545 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.023766 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.037328 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.097178 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-scripts\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.097248 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-run-httpd\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.097341 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.097366 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.097459 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.097494 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-config-data\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.097515 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvhx5\" (UniqueName: \"kubernetes.io/projected/561f373f-2c49-42ba-b353-721505b7d1e1-kube-api-access-pvhx5\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.097650 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-log-httpd\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.125840 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.199550 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-nova-metadata-tls-certs\") pod \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.199595 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-combined-ca-bundle\") pod \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.199740 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr6cw\" (UniqueName: \"kubernetes.io/projected/d6f1297d-f47b-4ac3-a9ba-3466972af97e-kube-api-access-sr6cw\") pod \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.199782 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-config-data\") pod \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.199878 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f1297d-f47b-4ac3-a9ba-3466972af97e-logs\") pod \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\" (UID: \"d6f1297d-f47b-4ac3-a9ba-3466972af97e\") " Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.200170 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.200210 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-config-data\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.200899 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6f1297d-f47b-4ac3-a9ba-3466972af97e-logs" (OuterVolumeSpecName: "logs") pod "d6f1297d-f47b-4ac3-a9ba-3466972af97e" (UID: "d6f1297d-f47b-4ac3-a9ba-3466972af97e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.201903 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvhx5\" (UniqueName: \"kubernetes.io/projected/561f373f-2c49-42ba-b353-721505b7d1e1-kube-api-access-pvhx5\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.201988 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-log-httpd\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.202010 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-scripts\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.202069 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-run-httpd\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.202167 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.202186 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.202301 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f1297d-f47b-4ac3-a9ba-3466972af97e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.208575 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f1297d-f47b-4ac3-a9ba-3466972af97e-kube-api-access-sr6cw" (OuterVolumeSpecName: "kube-api-access-sr6cw") pod "d6f1297d-f47b-4ac3-a9ba-3466972af97e" (UID: "d6f1297d-f47b-4ac3-a9ba-3466972af97e"). InnerVolumeSpecName "kube-api-access-sr6cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.208944 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-run-httpd\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.208967 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-log-httpd\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.213126 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d6f1297d-f47b-4ac3-a9ba-3466972af97e" (UID: "d6f1297d-f47b-4ac3-a9ba-3466972af97e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.216191 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.219022 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-config-data" (OuterVolumeSpecName: "config-data") pod "d6f1297d-f47b-4ac3-a9ba-3466972af97e" (UID: "d6f1297d-f47b-4ac3-a9ba-3466972af97e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.223828 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-scripts\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.230287 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6f1297d-f47b-4ac3-a9ba-3466972af97e" (UID: "d6f1297d-f47b-4ac3-a9ba-3466972af97e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.230883 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.234364 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.234424 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvhx5\" (UniqueName: \"kubernetes.io/projected/561f373f-2c49-42ba-b353-721505b7d1e1-kube-api-access-pvhx5\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.235267 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-config-data\") pod \"ceilometer-0\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.304533 4573 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.304817 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.304828 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr6cw\" (UniqueName: \"kubernetes.io/projected/d6f1297d-f47b-4ac3-a9ba-3466972af97e-kube-api-access-sr6cw\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.304838 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f1297d-f47b-4ac3-a9ba-3466972af97e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.393698 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.412287 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.930108 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:05 crc kubenswrapper[4573]: W1203 09:04:05.934380 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod561f373f_2c49_42ba_b353_721505b7d1e1.slice/crio-801614fe2a78bd7c10bdf77dd84753cef3d83d72b76c88007292fc3a142a30a5 WatchSource:0}: Error finding container 801614fe2a78bd7c10bdf77dd84753cef3d83d72b76c88007292fc3a142a30a5: Status 404 returned error can't find the container with id 801614fe2a78bd7c10bdf77dd84753cef3d83d72b76c88007292fc3a142a30a5 Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.948794 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e50ae89c-c3c5-44a5-9222-b14ddbadc294","Type":"ContainerStarted","Data":"1057d230a2433b7cb66b2a6e558e3b7de2602b7b2c4fee0b3c50f378b0a6024a"} Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.948945 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e50ae89c-c3c5-44a5-9222-b14ddbadc294","Type":"ContainerStarted","Data":"d9b425493633b4addb2eed68086d55ae033d8afa62a5391a8b6ec56449bd84de"} Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.948962 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.951586 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2227cd84-de1e-4b88-8712-dbd6471e6c67","Type":"ContainerStarted","Data":"a09aa1eb3a6f3caa37ee0d6b786b5ca2d2de88d6c1a691d7db373a74f4e50c49"} Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.951632 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2227cd84-de1e-4b88-8712-dbd6471e6c67","Type":"ContainerStarted","Data":"e0df76c6d433563667276b2a65c6f067bf851a8c309b90c0525b11c158350a54"} Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.951734 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.953886 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerStarted","Data":"801614fe2a78bd7c10bdf77dd84753cef3d83d72b76c88007292fc3a142a30a5"} Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.954029 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.970467 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.577091394 podStartE2EDuration="2.970452075s" podCreationTimestamp="2025-12-03 09:04:03 +0000 UTC" firstStartedPulling="2025-12-03 09:04:05.138608132 +0000 UTC m=+1565.706987391" lastFinishedPulling="2025-12-03 09:04:05.531968813 +0000 UTC m=+1566.100348072" observedRunningTime="2025-12-03 09:04:05.968268316 +0000 UTC m=+1566.536647575" watchObservedRunningTime="2025-12-03 09:04:05.970452075 +0000 UTC m=+1566.538831334" Dec 03 09:04:05 crc kubenswrapper[4573]: I1203 09:04:05.995628 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.9956093560000001 podStartE2EDuration="1.995609356s" podCreationTimestamp="2025-12-03 09:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:05.991136064 +0000 UTC m=+1566.559515323" watchObservedRunningTime="2025-12-03 09:04:05.995609356 +0000 UTC m=+1566.563988615" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.055855 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="441a63f9-caa0-4f25-8def-194be8fd64a6" path="/var/lib/kubelet/pods/441a63f9-caa0-4f25-8def-194be8fd64a6/volumes" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.059779 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="811326ff-3ff6-4d68-9c4f-4c356f09d6b2" path="/var/lib/kubelet/pods/811326ff-3ff6-4d68-9c4f-4c356f09d6b2/volumes" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.060339 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8a62ff8-6759-4a8e-971c-c8ab8ce2561f" path="/var/lib/kubelet/pods/f8a62ff8-6759-4a8e-971c-c8ab8ce2561f/volumes" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.094956 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.112298 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.127561 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.129544 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.134756 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.137350 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.140870 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.222926 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e482326-82c9-4621-896b-dcd607248b36-logs\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.223266 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghnmn\" (UniqueName: \"kubernetes.io/projected/4e482326-82c9-4621-896b-dcd607248b36-kube-api-access-ghnmn\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.223333 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.223362 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-config-data\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.223384 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.326174 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.326231 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-config-data\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.326252 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.326316 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e482326-82c9-4621-896b-dcd607248b36-logs\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.326391 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghnmn\" (UniqueName: \"kubernetes.io/projected/4e482326-82c9-4621-896b-dcd607248b36-kube-api-access-ghnmn\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.327252 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e482326-82c9-4621-896b-dcd607248b36-logs\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.332846 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.332948 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-config-data\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.333232 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.356822 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghnmn\" (UniqueName: \"kubernetes.io/projected/4e482326-82c9-4621-896b-dcd607248b36-kube-api-access-ghnmn\") pod \"nova-metadata-0\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " pod="openstack/nova-metadata-0" Dec 03 09:04:06 crc kubenswrapper[4573]: I1203 09:04:06.459251 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:07 crc kubenswrapper[4573]: I1203 09:04:07.053256 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:07 crc kubenswrapper[4573]: W1203 09:04:07.054291 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e482326_82c9_4621_896b_dcd607248b36.slice/crio-13e1413baf1abfa15486d0b6ae9e55dbe409b81857fbeed7ac00af9f13a63eb9 WatchSource:0}: Error finding container 13e1413baf1abfa15486d0b6ae9e55dbe409b81857fbeed7ac00af9f13a63eb9: Status 404 returned error can't find the container with id 13e1413baf1abfa15486d0b6ae9e55dbe409b81857fbeed7ac00af9f13a63eb9 Dec 03 09:04:07 crc kubenswrapper[4573]: I1203 09:04:07.977963 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerStarted","Data":"3899a4c5e236d6ea1934983729f4515e3cf4a60b9ac29e1075bd352c0c28590e"} Dec 03 09:04:07 crc kubenswrapper[4573]: I1203 09:04:07.981560 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e482326-82c9-4621-896b-dcd607248b36","Type":"ContainerStarted","Data":"5d9d61a43c20e0d52583c62fc4d630055a60204687c14c1f2855cd493ef8b399"} Dec 03 09:04:07 crc kubenswrapper[4573]: I1203 09:04:07.981609 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e482326-82c9-4621-896b-dcd607248b36","Type":"ContainerStarted","Data":"5b179ffe542f3b6ae273d19064c9aa5e5c205e88e6a09449ab437659c8fb9c00"} Dec 03 09:04:07 crc kubenswrapper[4573]: I1203 09:04:07.981620 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e482326-82c9-4621-896b-dcd607248b36","Type":"ContainerStarted","Data":"13e1413baf1abfa15486d0b6ae9e55dbe409b81857fbeed7ac00af9f13a63eb9"} Dec 03 09:04:07 crc kubenswrapper[4573]: I1203 09:04:07.994691 4573 generic.go:334] "Generic (PLEG): container finished" podID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerID="b986037d887208c104d14de970e22d137e4a6987234e42c13309978841fe05cd" exitCode=0 Dec 03 09:04:07 crc kubenswrapper[4573]: I1203 09:04:07.994747 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b2c14a2a-018a-4190-8f7d-fc55004eaabb","Type":"ContainerDied","Data":"b986037d887208c104d14de970e22d137e4a6987234e42c13309978841fe05cd"} Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.022582 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.022560719 podStartE2EDuration="2.022560719s" podCreationTimestamp="2025-12-03 09:04:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:08.018111516 +0000 UTC m=+1568.586490775" watchObservedRunningTime="2025-12-03 09:04:08.022560719 +0000 UTC m=+1568.590939978" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.056918 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f1297d-f47b-4ac3-a9ba-3466972af97e" path="/var/lib/kubelet/pods/d6f1297d-f47b-4ac3-a9ba-3466972af97e/volumes" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.322613 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.494640 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hb2x\" (UniqueName: \"kubernetes.io/projected/b2c14a2a-018a-4190-8f7d-fc55004eaabb-kube-api-access-7hb2x\") pod \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.495061 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-config-data\") pod \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.495185 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-combined-ca-bundle\") pod \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.495308 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2c14a2a-018a-4190-8f7d-fc55004eaabb-logs\") pod \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\" (UID: \"b2c14a2a-018a-4190-8f7d-fc55004eaabb\") " Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.495724 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2c14a2a-018a-4190-8f7d-fc55004eaabb-logs" (OuterVolumeSpecName: "logs") pod "b2c14a2a-018a-4190-8f7d-fc55004eaabb" (UID: "b2c14a2a-018a-4190-8f7d-fc55004eaabb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.495958 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2c14a2a-018a-4190-8f7d-fc55004eaabb-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.501935 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2c14a2a-018a-4190-8f7d-fc55004eaabb-kube-api-access-7hb2x" (OuterVolumeSpecName: "kube-api-access-7hb2x") pod "b2c14a2a-018a-4190-8f7d-fc55004eaabb" (UID: "b2c14a2a-018a-4190-8f7d-fc55004eaabb"). InnerVolumeSpecName "kube-api-access-7hb2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.537376 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-config-data" (OuterVolumeSpecName: "config-data") pod "b2c14a2a-018a-4190-8f7d-fc55004eaabb" (UID: "b2c14a2a-018a-4190-8f7d-fc55004eaabb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.549269 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2c14a2a-018a-4190-8f7d-fc55004eaabb" (UID: "b2c14a2a-018a-4190-8f7d-fc55004eaabb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.598182 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hb2x\" (UniqueName: \"kubernetes.io/projected/b2c14a2a-018a-4190-8f7d-fc55004eaabb-kube-api-access-7hb2x\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.598227 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.598241 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2c14a2a-018a-4190-8f7d-fc55004eaabb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:08 crc kubenswrapper[4573]: E1203 09:04:08.788830 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5db22f7d546abe1fa78ac7a4427f78a6d3425c8edb1499b37ee171da9e8b10f1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:08 crc kubenswrapper[4573]: E1203 09:04:08.791476 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5db22f7d546abe1fa78ac7a4427f78a6d3425c8edb1499b37ee171da9e8b10f1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:08 crc kubenswrapper[4573]: E1203 09:04:08.794816 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5db22f7d546abe1fa78ac7a4427f78a6d3425c8edb1499b37ee171da9e8b10f1" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:08 crc kubenswrapper[4573]: E1203 09:04:08.794856 4573 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="01ed2237-985e-4599-b585-842558702f87" containerName="nova-scheduler-scheduler" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.873368 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:04:08 crc kubenswrapper[4573]: I1203 09:04:08.873433 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.007575 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b2c14a2a-018a-4190-8f7d-fc55004eaabb","Type":"ContainerDied","Data":"0fde18d388c8fe5484f4d103ac4b27ac22f34b1708705af832fa0cfd8b846fa9"} Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.007637 4573 scope.go:117] "RemoveContainer" containerID="b986037d887208c104d14de970e22d137e4a6987234e42c13309978841fe05cd" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.007829 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.010950 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerStarted","Data":"73fb1ec2d0e4d00edcefaae9cd750ae6ffa54046dcd3990daa3e62cbf311bb66"} Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.042682 4573 scope.go:117] "RemoveContainer" containerID="47588117147ca6a44198ed5f34f84e21d454adfd06e8b910348af22ec4065836" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.078033 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.090110 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.107948 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:09 crc kubenswrapper[4573]: E1203 09:04:09.108507 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-api" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.108528 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-api" Dec 03 09:04:09 crc kubenswrapper[4573]: E1203 09:04:09.108554 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-log" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.108560 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-log" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.108795 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-log" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.108824 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" containerName="nova-api-api" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.110038 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.111828 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.115573 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.212931 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.213365 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-config-data\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.213633 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-logs\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.213735 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-962jb\" (UniqueName: \"kubernetes.io/projected/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-kube-api-access-962jb\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.316322 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-config-data\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.316735 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-logs\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.316807 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-962jb\" (UniqueName: \"kubernetes.io/projected/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-kube-api-access-962jb\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.316876 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.317721 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-logs\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.320686 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.327684 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-config-data\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.346690 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-962jb\" (UniqueName: \"kubernetes.io/projected/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-kube-api-access-962jb\") pod \"nova-api-0\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.502671 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:09 crc kubenswrapper[4573]: I1203 09:04:09.938786 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-xrv4g" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="registry-server" probeResult="failure" output=< Dec 03 09:04:09 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:04:09 crc kubenswrapper[4573]: > Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.000135 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.090145 4573 generic.go:334] "Generic (PLEG): container finished" podID="01ed2237-985e-4599-b585-842558702f87" containerID="5db22f7d546abe1fa78ac7a4427f78a6d3425c8edb1499b37ee171da9e8b10f1" exitCode=0 Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.093864 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2c14a2a-018a-4190-8f7d-fc55004eaabb" path="/var/lib/kubelet/pods/b2c14a2a-018a-4190-8f7d-fc55004eaabb/volumes" Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.095031 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cda7757-bb11-46f2-9bac-a4f81a66bc1c","Type":"ContainerStarted","Data":"b8a7c5b242778a8baac62d9c1be38cb8dc99f1d61ccf5a503aa679dd04f1e15f"} Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.095091 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"01ed2237-985e-4599-b585-842558702f87","Type":"ContainerDied","Data":"5db22f7d546abe1fa78ac7a4427f78a6d3425c8edb1499b37ee171da9e8b10f1"} Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.095241 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerStarted","Data":"dc2588fd72227607c51a6e9d6ad483de1512389881979131ffb2a347fa23eb94"} Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.241875 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.343579 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-combined-ca-bundle\") pod \"01ed2237-985e-4599-b585-842558702f87\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.343678 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hldm2\" (UniqueName: \"kubernetes.io/projected/01ed2237-985e-4599-b585-842558702f87-kube-api-access-hldm2\") pod \"01ed2237-985e-4599-b585-842558702f87\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.343851 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-config-data\") pod \"01ed2237-985e-4599-b585-842558702f87\" (UID: \"01ed2237-985e-4599-b585-842558702f87\") " Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.359460 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ed2237-985e-4599-b585-842558702f87-kube-api-access-hldm2" (OuterVolumeSpecName: "kube-api-access-hldm2") pod "01ed2237-985e-4599-b585-842558702f87" (UID: "01ed2237-985e-4599-b585-842558702f87"). InnerVolumeSpecName "kube-api-access-hldm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.430660 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-config-data" (OuterVolumeSpecName: "config-data") pod "01ed2237-985e-4599-b585-842558702f87" (UID: "01ed2237-985e-4599-b585-842558702f87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.443614 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01ed2237-985e-4599-b585-842558702f87" (UID: "01ed2237-985e-4599-b585-842558702f87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.446118 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.446145 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hldm2\" (UniqueName: \"kubernetes.io/projected/01ed2237-985e-4599-b585-842558702f87-kube-api-access-hldm2\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:10 crc kubenswrapper[4573]: I1203 09:04:10.446161 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01ed2237-985e-4599-b585-842558702f87-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.108959 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerStarted","Data":"6aaf34b1cdeb041f67fc27c7f3a8f073f63ade7829322ee3e0adfe059ddea95a"} Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.121205 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cda7757-bb11-46f2-9bac-a4f81a66bc1c","Type":"ContainerStarted","Data":"f4b04087942e156ae4f9e1a0f6cdc6f5106054f34667179e1ca85ccd2dd2eb15"} Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.121277 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cda7757-bb11-46f2-9bac-a4f81a66bc1c","Type":"ContainerStarted","Data":"07e99f411f81ebd1c7d913a96bd39aa4350bc0201818a0cec99f71426fa35ab6"} Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.126539 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"01ed2237-985e-4599-b585-842558702f87","Type":"ContainerDied","Data":"fcbd683d64ceb76b994a8cda2e39cc907891738a785863bb7d6e373b40badd67"} Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.126616 4573 scope.go:117] "RemoveContainer" containerID="5db22f7d546abe1fa78ac7a4427f78a6d3425c8edb1499b37ee171da9e8b10f1" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.126655 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.163190 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.865443364 podStartE2EDuration="7.163171301s" podCreationTimestamp="2025-12-03 09:04:04 +0000 UTC" firstStartedPulling="2025-12-03 09:04:05.936498692 +0000 UTC m=+1566.504877941" lastFinishedPulling="2025-12-03 09:04:10.234226619 +0000 UTC m=+1570.802605878" observedRunningTime="2025-12-03 09:04:11.134833952 +0000 UTC m=+1571.703213211" watchObservedRunningTime="2025-12-03 09:04:11.163171301 +0000 UTC m=+1571.731550560" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.175397 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.175381586 podStartE2EDuration="2.175381586s" podCreationTimestamp="2025-12-03 09:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:11.157955187 +0000 UTC m=+1571.726334446" watchObservedRunningTime="2025-12-03 09:04:11.175381586 +0000 UTC m=+1571.743760835" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.227348 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.238967 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.263369 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:11 crc kubenswrapper[4573]: E1203 09:04:11.264123 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ed2237-985e-4599-b585-842558702f87" containerName="nova-scheduler-scheduler" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.264285 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ed2237-985e-4599-b585-842558702f87" containerName="nova-scheduler-scheduler" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.264549 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ed2237-985e-4599-b585-842558702f87" containerName="nova-scheduler-scheduler" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.265800 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.268322 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.292122 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.364418 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.364492 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-config-data\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.364517 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5x7f\" (UniqueName: \"kubernetes.io/projected/70534db4-de8c-4c04-92d6-60abef7c4843-kube-api-access-s5x7f\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.460329 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.461547 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.465559 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.465620 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-config-data\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.465643 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5x7f\" (UniqueName: \"kubernetes.io/projected/70534db4-de8c-4c04-92d6-60abef7c4843-kube-api-access-s5x7f\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.470358 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-config-data\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.470840 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.481334 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5x7f\" (UniqueName: \"kubernetes.io/projected/70534db4-de8c-4c04-92d6-60abef7c4843-kube-api-access-s5x7f\") pod \"nova-scheduler-0\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:11 crc kubenswrapper[4573]: I1203 09:04:11.593943 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:04:12 crc kubenswrapper[4573]: I1203 09:04:12.045440 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ed2237-985e-4599-b585-842558702f87" path="/var/lib/kubelet/pods/01ed2237-985e-4599-b585-842558702f87/volumes" Dec 03 09:04:12 crc kubenswrapper[4573]: I1203 09:04:12.137560 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:04:12 crc kubenswrapper[4573]: I1203 09:04:12.163835 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:13 crc kubenswrapper[4573]: I1203 09:04:13.148648 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70534db4-de8c-4c04-92d6-60abef7c4843","Type":"ContainerStarted","Data":"9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535"} Dec 03 09:04:13 crc kubenswrapper[4573]: I1203 09:04:13.148916 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70534db4-de8c-4c04-92d6-60abef7c4843","Type":"ContainerStarted","Data":"c4900c7ee1ca64fa9598cf359330da86f3079491c4f423096efad4616a4f69b5"} Dec 03 09:04:14 crc kubenswrapper[4573]: I1203 09:04:14.604216 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 09:04:14 crc kubenswrapper[4573]: I1203 09:04:14.623923 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.623881131 podStartE2EDuration="3.623881131s" podCreationTimestamp="2025-12-03 09:04:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:13.164489579 +0000 UTC m=+1573.732868838" watchObservedRunningTime="2025-12-03 09:04:14.623881131 +0000 UTC m=+1575.192260390" Dec 03 09:04:14 crc kubenswrapper[4573]: I1203 09:04:14.757617 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 09:04:16 crc kubenswrapper[4573]: I1203 09:04:16.181116 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmzcx" event={"ID":"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb","Type":"ContainerStarted","Data":"d5fadf69c2bdc0529e7c65e6997322558549097bb56b7fc8decdb205a57c7093"} Dec 03 09:04:16 crc kubenswrapper[4573]: I1203 09:04:16.460372 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 09:04:16 crc kubenswrapper[4573]: I1203 09:04:16.460435 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 09:04:16 crc kubenswrapper[4573]: I1203 09:04:16.594238 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 09:04:17 crc kubenswrapper[4573]: I1203 09:04:17.477269 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:04:17 crc kubenswrapper[4573]: I1203 09:04:17.477255 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:04:18 crc kubenswrapper[4573]: I1203 09:04:18.935564 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:18.999962 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.218011 4573 generic.go:334] "Generic (PLEG): container finished" podID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerID="39cb62ba4ad569a45049cde16bc1cf40005dce1875fe10337e5b5ae8bcb634a7" exitCode=137 Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.218959 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerDied","Data":"39cb62ba4ad569a45049cde16bc1cf40005dce1875fe10337e5b5ae8bcb634a7"} Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.503812 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.504208 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.641729 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xrv4g"] Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.787522 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.844632 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c69ceb8-3d0e-4920-be28-36799e264f96-logs\") pod \"8c69ceb8-3d0e-4920-be28-36799e264f96\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.844783 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-secret-key\") pod \"8c69ceb8-3d0e-4920-be28-36799e264f96\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.844818 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-combined-ca-bundle\") pod \"8c69ceb8-3d0e-4920-be28-36799e264f96\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.844882 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-config-data\") pod \"8c69ceb8-3d0e-4920-be28-36799e264f96\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.844945 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-scripts\") pod \"8c69ceb8-3d0e-4920-be28-36799e264f96\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.844995 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-577vk\" (UniqueName: \"kubernetes.io/projected/8c69ceb8-3d0e-4920-be28-36799e264f96-kube-api-access-577vk\") pod \"8c69ceb8-3d0e-4920-be28-36799e264f96\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.845029 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-tls-certs\") pod \"8c69ceb8-3d0e-4920-be28-36799e264f96\" (UID: \"8c69ceb8-3d0e-4920-be28-36799e264f96\") " Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.845682 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c69ceb8-3d0e-4920-be28-36799e264f96-logs" (OuterVolumeSpecName: "logs") pod "8c69ceb8-3d0e-4920-be28-36799e264f96" (UID: "8c69ceb8-3d0e-4920-be28-36799e264f96"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.899226 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8c69ceb8-3d0e-4920-be28-36799e264f96" (UID: "8c69ceb8-3d0e-4920-be28-36799e264f96"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.915364 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-config-data" (OuterVolumeSpecName: "config-data") pod "8c69ceb8-3d0e-4920-be28-36799e264f96" (UID: "8c69ceb8-3d0e-4920-be28-36799e264f96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.925238 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c69ceb8-3d0e-4920-be28-36799e264f96-kube-api-access-577vk" (OuterVolumeSpecName: "kube-api-access-577vk") pod "8c69ceb8-3d0e-4920-be28-36799e264f96" (UID: "8c69ceb8-3d0e-4920-be28-36799e264f96"). InnerVolumeSpecName "kube-api-access-577vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.932941 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c69ceb8-3d0e-4920-be28-36799e264f96" (UID: "8c69ceb8-3d0e-4920-be28-36799e264f96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.935681 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-scripts" (OuterVolumeSpecName: "scripts") pod "8c69ceb8-3d0e-4920-be28-36799e264f96" (UID: "8c69ceb8-3d0e-4920-be28-36799e264f96"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.948609 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c69ceb8-3d0e-4920-be28-36799e264f96-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.948640 4573 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.948651 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.948661 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.948672 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c69ceb8-3d0e-4920-be28-36799e264f96-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:19 crc kubenswrapper[4573]: I1203 09:04:19.948681 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-577vk\" (UniqueName: \"kubernetes.io/projected/8c69ceb8-3d0e-4920-be28-36799e264f96-kube-api-access-577vk\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.006390 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "8c69ceb8-3d0e-4920-be28-36799e264f96" (UID: "8c69ceb8-3d0e-4920-be28-36799e264f96"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.050794 4573 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c69ceb8-3d0e-4920-be28-36799e264f96-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.249166 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b5b6556-4zxxs" event={"ID":"8c69ceb8-3d0e-4920-be28-36799e264f96","Type":"ContainerDied","Data":"4dc9c0026737b9e06e10ab08990fb2f6fac86fe8064e66cd1d53ce7d88f5fd13"} Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.249253 4573 scope.go:117] "RemoveContainer" containerID="d0c803805d4bd3808e5ee89977579fff41f3ca9da62cdfa598f77cba0d2c5e22" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.249539 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8b5b6556-4zxxs" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.266706 4573 generic.go:334] "Generic (PLEG): container finished" podID="142ca013-7220-454f-9e00-05af4f3f0ded" containerID="5babce96efa2d3c058b8fc06692f93399e8e3ae4be8fda57861915397e3d9bc6" exitCode=137 Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.266822 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"142ca013-7220-454f-9e00-05af4f3f0ded","Type":"ContainerDied","Data":"5babce96efa2d3c058b8fc06692f93399e8e3ae4be8fda57861915397e3d9bc6"} Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.272764 4573 generic.go:334] "Generic (PLEG): container finished" podID="f4d1c7fd-3939-4391-ae35-24bc7c2beaeb" containerID="d5fadf69c2bdc0529e7c65e6997322558549097bb56b7fc8decdb205a57c7093" exitCode=0 Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.272837 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmzcx" event={"ID":"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb","Type":"ContainerDied","Data":"d5fadf69c2bdc0529e7c65e6997322558549097bb56b7fc8decdb205a57c7093"} Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.273296 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xrv4g" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="registry-server" containerID="cri-o://b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4" gracePeriod=2 Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.347882 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b8b5b6556-4zxxs"] Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.359017 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7b8b5b6556-4zxxs"] Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.535992 4573 scope.go:117] "RemoveContainer" containerID="39cb62ba4ad569a45049cde16bc1cf40005dce1875fe10337e5b5ae8bcb634a7" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.592927 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.200:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.593534 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.200:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.711350 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.792649 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q887\" (UniqueName: \"kubernetes.io/projected/142ca013-7220-454f-9e00-05af4f3f0ded-kube-api-access-2q887\") pod \"142ca013-7220-454f-9e00-05af4f3f0ded\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.792797 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-combined-ca-bundle\") pod \"142ca013-7220-454f-9e00-05af4f3f0ded\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.792950 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-config-data\") pod \"142ca013-7220-454f-9e00-05af4f3f0ded\" (UID: \"142ca013-7220-454f-9e00-05af4f3f0ded\") " Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.813874 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/142ca013-7220-454f-9e00-05af4f3f0ded-kube-api-access-2q887" (OuterVolumeSpecName: "kube-api-access-2q887") pod "142ca013-7220-454f-9e00-05af4f3f0ded" (UID: "142ca013-7220-454f-9e00-05af4f3f0ded"). InnerVolumeSpecName "kube-api-access-2q887". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.876572 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "142ca013-7220-454f-9e00-05af4f3f0ded" (UID: "142ca013-7220-454f-9e00-05af4f3f0ded"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.905285 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q887\" (UniqueName: \"kubernetes.io/projected/142ca013-7220-454f-9e00-05af4f3f0ded-kube-api-access-2q887\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.905327 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.939936 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-config-data" (OuterVolumeSpecName: "config-data") pod "142ca013-7220-454f-9e00-05af4f3f0ded" (UID: "142ca013-7220-454f-9e00-05af4f3f0ded"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:20 crc kubenswrapper[4573]: I1203 09:04:20.942550 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.007187 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/142ca013-7220-454f-9e00-05af4f3f0ded-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.111797 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-utilities\") pod \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.111872 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf45v\" (UniqueName: \"kubernetes.io/projected/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-kube-api-access-zf45v\") pod \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.112072 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-catalog-content\") pod \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\" (UID: \"f48f1b5a-0f0f-4491-bab1-6b602cfefb03\") " Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.112448 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-utilities" (OuterVolumeSpecName: "utilities") pod "f48f1b5a-0f0f-4491-bab1-6b602cfefb03" (UID: "f48f1b5a-0f0f-4491-bab1-6b602cfefb03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.112759 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.117298 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-kube-api-access-zf45v" (OuterVolumeSpecName: "kube-api-access-zf45v") pod "f48f1b5a-0f0f-4491-bab1-6b602cfefb03" (UID: "f48f1b5a-0f0f-4491-bab1-6b602cfefb03"). InnerVolumeSpecName "kube-api-access-zf45v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.168180 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f48f1b5a-0f0f-4491-bab1-6b602cfefb03" (UID: "f48f1b5a-0f0f-4491-bab1-6b602cfefb03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.214494 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf45v\" (UniqueName: \"kubernetes.io/projected/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-kube-api-access-zf45v\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.214534 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48f1b5a-0f0f-4491-bab1-6b602cfefb03-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.283517 4573 generic.go:334] "Generic (PLEG): container finished" podID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerID="b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4" exitCode=0 Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.283584 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrv4g" event={"ID":"f48f1b5a-0f0f-4491-bab1-6b602cfefb03","Type":"ContainerDied","Data":"b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4"} Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.283609 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xrv4g" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.283613 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xrv4g" event={"ID":"f48f1b5a-0f0f-4491-bab1-6b602cfefb03","Type":"ContainerDied","Data":"084daedbd5c9ce61014f6776c985169279eb4bc5f18dab29823b9875e42cc4a8"} Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.283627 4573 scope.go:117] "RemoveContainer" containerID="b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.287224 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmzcx" event={"ID":"f4d1c7fd-3939-4391-ae35-24bc7c2beaeb","Type":"ContainerStarted","Data":"b5be75fe38e375ecc566e0beaddf63326b8b0c62d5c496d89f360d04267d29a5"} Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.293212 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"142ca013-7220-454f-9e00-05af4f3f0ded","Type":"ContainerDied","Data":"758c82ba9a0c8f8bacbd96c0a18c906fb74398de713d977bfe8987a062235c0e"} Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.293289 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.306776 4573 scope.go:117] "RemoveContainer" containerID="076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.329904 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tmzcx" podStartSLOduration=2.23975394 podStartE2EDuration="48.32988548s" podCreationTimestamp="2025-12-03 09:03:33 +0000 UTC" firstStartedPulling="2025-12-03 09:03:34.679132691 +0000 UTC m=+1535.247511950" lastFinishedPulling="2025-12-03 09:04:20.769264231 +0000 UTC m=+1581.337643490" observedRunningTime="2025-12-03 09:04:21.321803808 +0000 UTC m=+1581.890183067" watchObservedRunningTime="2025-12-03 09:04:21.32988548 +0000 UTC m=+1581.898264739" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.348664 4573 scope.go:117] "RemoveContainer" containerID="83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.374919 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.424389 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.425232 4573 scope.go:117] "RemoveContainer" containerID="b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.425836 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4\": container with ID starting with b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4 not found: ID does not exist" containerID="b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.425897 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4"} err="failed to get container status \"b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4\": rpc error: code = NotFound desc = could not find container \"b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4\": container with ID starting with b2efb8e2cb45c28bdb6f2015045cad25fc82c03b977f2a11d2ab3225010803b4 not found: ID does not exist" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.425929 4573 scope.go:117] "RemoveContainer" containerID="076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.431251 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af\": container with ID starting with 076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af not found: ID does not exist" containerID="076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.431334 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af"} err="failed to get container status \"076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af\": rpc error: code = NotFound desc = could not find container \"076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af\": container with ID starting with 076f24aed91510be90b1bde3469bc9ce3f2f3d1226c343d427ec8bbb86edb1af not found: ID does not exist" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.431378 4573 scope.go:117] "RemoveContainer" containerID="83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.431979 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53\": container with ID starting with 83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53 not found: ID does not exist" containerID="83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.432037 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53"} err="failed to get container status \"83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53\": rpc error: code = NotFound desc = could not find container \"83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53\": container with ID starting with 83aea83595b92069cedc2e128624cd10ffb8d58c0ee0d17ec8a06ce598302b53 not found: ID does not exist" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.432092 4573 scope.go:117] "RemoveContainer" containerID="5babce96efa2d3c058b8fc06692f93399e8e3ae4be8fda57861915397e3d9bc6" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.451961 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xrv4g"] Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.464112 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xrv4g"] Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.477937 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.478727 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.478749 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.478763 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="extract-content" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.478771 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="extract-content" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.478815 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="extract-utilities" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.478823 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="extract-utilities" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.478836 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="registry-server" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.478843 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="registry-server" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.478870 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon-log" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.478877 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon-log" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.478890 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.478896 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.478905 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="142ca013-7220-454f-9e00-05af4f3f0ded" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.478912 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="142ca013-7220-454f-9e00-05af4f3f0ded" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 09:04:21 crc kubenswrapper[4573]: E1203 09:04:21.478924 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.478931 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.479406 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.479456 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon-log" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.479474 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="142ca013-7220-454f-9e00-05af4f3f0ded" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.479497 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" containerName="registry-server" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.479506 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.481669 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.489853 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.490208 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.490388 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.516012 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.594084 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.626130 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.626196 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.626289 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.626339 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.626378 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtbjm\" (UniqueName: \"kubernetes.io/projected/192989bd-f8d0-43d2-9267-3022e451041c-kube-api-access-gtbjm\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.650778 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.731979 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.732057 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtbjm\" (UniqueName: \"kubernetes.io/projected/192989bd-f8d0-43d2-9267-3022e451041c-kube-api-access-gtbjm\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.732119 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.732147 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.732602 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.739299 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.745877 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.754444 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.755543 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/192989bd-f8d0-43d2-9267-3022e451041c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.756855 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtbjm\" (UniqueName: \"kubernetes.io/projected/192989bd-f8d0-43d2-9267-3022e451041c-kube-api-access-gtbjm\") pod \"nova-cell1-novncproxy-0\" (UID: \"192989bd-f8d0-43d2-9267-3022e451041c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:21 crc kubenswrapper[4573]: I1203 09:04:21.816615 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:22 crc kubenswrapper[4573]: I1203 09:04:22.051003 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="142ca013-7220-454f-9e00-05af4f3f0ded" path="/var/lib/kubelet/pods/142ca013-7220-454f-9e00-05af4f3f0ded/volumes" Dec 03 09:04:22 crc kubenswrapper[4573]: I1203 09:04:22.052098 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" path="/var/lib/kubelet/pods/8c69ceb8-3d0e-4920-be28-36799e264f96/volumes" Dec 03 09:04:22 crc kubenswrapper[4573]: I1203 09:04:22.055586 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f48f1b5a-0f0f-4491-bab1-6b602cfefb03" path="/var/lib/kubelet/pods/f48f1b5a-0f0f-4491-bab1-6b602cfefb03/volumes" Dec 03 09:04:22 crc kubenswrapper[4573]: I1203 09:04:22.331379 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 09:04:22 crc kubenswrapper[4573]: I1203 09:04:22.402146 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 09:04:23 crc kubenswrapper[4573]: I1203 09:04:23.325387 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"192989bd-f8d0-43d2-9267-3022e451041c","Type":"ContainerStarted","Data":"3a7e643f31f1de0585891fa92702a66d53c580b6dffa7606dc893a76fdd050d3"} Dec 03 09:04:23 crc kubenswrapper[4573]: I1203 09:04:23.325591 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"192989bd-f8d0-43d2-9267-3022e451041c","Type":"ContainerStarted","Data":"60d5c19dbac8f1cbb2caba9a1f2bab86d845893d05ab1143ec5bca2ba978cb07"} Dec 03 09:04:23 crc kubenswrapper[4573]: I1203 09:04:23.359753 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.359731713 podStartE2EDuration="2.359731713s" podCreationTimestamp="2025-12-03 09:04:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:23.344334069 +0000 UTC m=+1583.912713328" watchObservedRunningTime="2025-12-03 09:04:23.359731713 +0000 UTC m=+1583.928110972" Dec 03 09:04:23 crc kubenswrapper[4573]: I1203 09:04:23.635263 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:04:23 crc kubenswrapper[4573]: I1203 09:04:23.635569 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:04:24 crc kubenswrapper[4573]: I1203 09:04:24.713802 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tmzcx" podUID="f4d1c7fd-3939-4391-ae35-24bc7c2beaeb" containerName="registry-server" probeResult="failure" output=< Dec 03 09:04:24 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:04:24 crc kubenswrapper[4573]: > Dec 03 09:04:26 crc kubenswrapper[4573]: I1203 09:04:26.465141 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 09:04:26 crc kubenswrapper[4573]: I1203 09:04:26.468989 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 09:04:26 crc kubenswrapper[4573]: I1203 09:04:26.476111 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 09:04:26 crc kubenswrapper[4573]: I1203 09:04:26.817217 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:27 crc kubenswrapper[4573]: I1203 09:04:27.366215 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.508741 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.509185 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.509399 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.509479 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.514106 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.519498 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.825441 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-h955r"] Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.826130 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c69ceb8-3d0e-4920-be28-36799e264f96" containerName="horizon" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.833523 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.909239 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-h955r"] Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.922857 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.922927 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2ptv\" (UniqueName: \"kubernetes.io/projected/4082d0f4-8d86-4382-bcb6-fe560d365ae3-kube-api-access-p2ptv\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.922969 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-config\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.922995 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.923082 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:29 crc kubenswrapper[4573]: I1203 09:04:29.923110 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.027186 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.027293 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.027344 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2ptv\" (UniqueName: \"kubernetes.io/projected/4082d0f4-8d86-4382-bcb6-fe560d365ae3-kube-api-access-p2ptv\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.027386 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-config\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.027411 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.027495 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.028391 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.028804 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.029276 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.040830 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-config\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.070409 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.082743 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2ptv\" (UniqueName: \"kubernetes.io/projected/4082d0f4-8d86-4382-bcb6-fe560d365ae3-kube-api-access-p2ptv\") pod \"dnsmasq-dns-cd5cbd7b9-h955r\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.206587 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:30 crc kubenswrapper[4573]: I1203 09:04:30.761123 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-h955r"] Dec 03 09:04:31 crc kubenswrapper[4573]: I1203 09:04:31.408821 4573 generic.go:334] "Generic (PLEG): container finished" podID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" containerID="a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35" exitCode=0 Dec 03 09:04:31 crc kubenswrapper[4573]: I1203 09:04:31.408877 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" event={"ID":"4082d0f4-8d86-4382-bcb6-fe560d365ae3","Type":"ContainerDied","Data":"a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35"} Dec 03 09:04:31 crc kubenswrapper[4573]: I1203 09:04:31.409422 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" event={"ID":"4082d0f4-8d86-4382-bcb6-fe560d365ae3","Type":"ContainerStarted","Data":"9ade5d4d4435801d78cf2a099c8f6ce13117fec395ce8458908e3b2c98c464b0"} Dec 03 09:04:31 crc kubenswrapper[4573]: I1203 09:04:31.817596 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:31 crc kubenswrapper[4573]: I1203 09:04:31.898632 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.426220 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" event={"ID":"4082d0f4-8d86-4382-bcb6-fe560d365ae3","Type":"ContainerStarted","Data":"64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e"} Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.426292 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.460241 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" podStartSLOduration=3.460215985 podStartE2EDuration="3.460215985s" podCreationTimestamp="2025-12-03 09:04:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:32.451199837 +0000 UTC m=+1593.019579126" watchObservedRunningTime="2025-12-03 09:04:32.460215985 +0000 UTC m=+1593.028595244" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.460869 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.764474 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-pgtv2"] Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.765796 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.784782 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pgtv2"] Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.788650 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.788674 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.902885 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-config-data\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.902944 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.902970 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59gpj\" (UniqueName: \"kubernetes.io/projected/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-kube-api-access-59gpj\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.903010 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-scripts\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.903071 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.903234 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-log" containerID="cri-o://07e99f411f81ebd1c7d913a96bd39aa4350bc0201818a0cec99f71426fa35ab6" gracePeriod=30 Dec 03 09:04:32 crc kubenswrapper[4573]: I1203 09:04:32.903427 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-api" containerID="cri-o://f4b04087942e156ae4f9e1a0f6cdc6f5106054f34667179e1ca85ccd2dd2eb15" gracePeriod=30 Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.005421 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.005585 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-config-data\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.005633 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59gpj\" (UniqueName: \"kubernetes.io/projected/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-kube-api-access-59gpj\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.005659 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-scripts\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.021305 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.026208 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-config-data\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.027571 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59gpj\" (UniqueName: \"kubernetes.io/projected/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-kube-api-access-59gpj\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.044029 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-scripts\") pod \"nova-cell1-cell-mapping-pgtv2\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.087198 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.445945 4573 generic.go:334] "Generic (PLEG): container finished" podID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerID="07e99f411f81ebd1c7d913a96bd39aa4350bc0201818a0cec99f71426fa35ab6" exitCode=143 Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.447436 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cda7757-bb11-46f2-9bac-a4f81a66bc1c","Type":"ContainerDied","Data":"07e99f411f81ebd1c7d913a96bd39aa4350bc0201818a0cec99f71426fa35ab6"} Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.723967 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-pgtv2"] Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.734905 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.738746 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="ceilometer-central-agent" containerID="cri-o://3899a4c5e236d6ea1934983729f4515e3cf4a60b9ac29e1075bd352c0c28590e" gracePeriod=30 Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.739508 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="proxy-httpd" containerID="cri-o://6aaf34b1cdeb041f67fc27c7f3a8f073f63ade7829322ee3e0adfe059ddea95a" gracePeriod=30 Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.739758 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="ceilometer-notification-agent" containerID="cri-o://73fb1ec2d0e4d00edcefaae9cd750ae6ffa54046dcd3990daa3e62cbf311bb66" gracePeriod=30 Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.739908 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="sg-core" containerID="cri-o://dc2588fd72227607c51a6e9d6ad483de1512389881979131ffb2a347fa23eb94" gracePeriod=30 Dec 03 09:04:33 crc kubenswrapper[4573]: I1203 09:04:33.755721 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.198:3000/\": EOF" Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.474696 4573 generic.go:334] "Generic (PLEG): container finished" podID="561f373f-2c49-42ba-b353-721505b7d1e1" containerID="6aaf34b1cdeb041f67fc27c7f3a8f073f63ade7829322ee3e0adfe059ddea95a" exitCode=0 Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.475067 4573 generic.go:334] "Generic (PLEG): container finished" podID="561f373f-2c49-42ba-b353-721505b7d1e1" containerID="dc2588fd72227607c51a6e9d6ad483de1512389881979131ffb2a347fa23eb94" exitCode=2 Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.475081 4573 generic.go:334] "Generic (PLEG): container finished" podID="561f373f-2c49-42ba-b353-721505b7d1e1" containerID="3899a4c5e236d6ea1934983729f4515e3cf4a60b9ac29e1075bd352c0c28590e" exitCode=0 Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.475136 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerDied","Data":"6aaf34b1cdeb041f67fc27c7f3a8f073f63ade7829322ee3e0adfe059ddea95a"} Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.475169 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerDied","Data":"dc2588fd72227607c51a6e9d6ad483de1512389881979131ffb2a347fa23eb94"} Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.475184 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerDied","Data":"3899a4c5e236d6ea1934983729f4515e3cf4a60b9ac29e1075bd352c0c28590e"} Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.477337 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pgtv2" event={"ID":"178dc07a-005e-4e3d-9c12-28b42eb2cc7d","Type":"ContainerStarted","Data":"04af37895750db5d16417f6b792e13afe02547877c30ce1dd17ba189134517bd"} Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.477368 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pgtv2" event={"ID":"178dc07a-005e-4e3d-9c12-28b42eb2cc7d","Type":"ContainerStarted","Data":"baf6a43b27f8c671b6eee22655f1fc9365ec2849fd04a9dd3db244e88cdc18bc"} Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.506238 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-pgtv2" podStartSLOduration=2.506222241 podStartE2EDuration="2.506222241s" podCreationTimestamp="2025-12-03 09:04:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:34.501393258 +0000 UTC m=+1595.069772517" watchObservedRunningTime="2025-12-03 09:04:34.506222241 +0000 UTC m=+1595.074601500" Dec 03 09:04:34 crc kubenswrapper[4573]: I1203 09:04:34.712118 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tmzcx" podUID="f4d1c7fd-3939-4391-ae35-24bc7c2beaeb" containerName="registry-server" probeResult="failure" output=< Dec 03 09:04:34 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:04:34 crc kubenswrapper[4573]: > Dec 03 09:04:35 crc kubenswrapper[4573]: I1203 09:04:35.395199 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.198:3000/\": dial tcp 10.217.0.198:3000: connect: connection refused" Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.502153 4573 generic.go:334] "Generic (PLEG): container finished" podID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerID="f4b04087942e156ae4f9e1a0f6cdc6f5106054f34667179e1ca85ccd2dd2eb15" exitCode=0 Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.502429 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cda7757-bb11-46f2-9bac-a4f81a66bc1c","Type":"ContainerDied","Data":"f4b04087942e156ae4f9e1a0f6cdc6f5106054f34667179e1ca85ccd2dd2eb15"} Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.765456 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.941523 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-logs\") pod \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.941680 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-config-data\") pod \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.941815 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-combined-ca-bundle\") pod \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.942318 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-logs" (OuterVolumeSpecName: "logs") pod "7cda7757-bb11-46f2-9bac-a4f81a66bc1c" (UID: "7cda7757-bb11-46f2-9bac-a4f81a66bc1c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.942904 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-962jb\" (UniqueName: \"kubernetes.io/projected/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-kube-api-access-962jb\") pod \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\" (UID: \"7cda7757-bb11-46f2-9bac-a4f81a66bc1c\") " Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.943423 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.969563 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-kube-api-access-962jb" (OuterVolumeSpecName: "kube-api-access-962jb") pod "7cda7757-bb11-46f2-9bac-a4f81a66bc1c" (UID: "7cda7757-bb11-46f2-9bac-a4f81a66bc1c"). InnerVolumeSpecName "kube-api-access-962jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:36 crc kubenswrapper[4573]: I1203 09:04:36.986607 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-config-data" (OuterVolumeSpecName: "config-data") pod "7cda7757-bb11-46f2-9bac-a4f81a66bc1c" (UID: "7cda7757-bb11-46f2-9bac-a4f81a66bc1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.016237 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cda7757-bb11-46f2-9bac-a4f81a66bc1c" (UID: "7cda7757-bb11-46f2-9bac-a4f81a66bc1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.051555 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.051598 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-962jb\" (UniqueName: \"kubernetes.io/projected/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-kube-api-access-962jb\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.051618 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cda7757-bb11-46f2-9bac-a4f81a66bc1c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.569784 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cda7757-bb11-46f2-9bac-a4f81a66bc1c","Type":"ContainerDied","Data":"b8a7c5b242778a8baac62d9c1be38cb8dc99f1d61ccf5a503aa679dd04f1e15f"} Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.569838 4573 scope.go:117] "RemoveContainer" containerID="f4b04087942e156ae4f9e1a0f6cdc6f5106054f34667179e1ca85ccd2dd2eb15" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.569995 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.583330 4573 generic.go:334] "Generic (PLEG): container finished" podID="561f373f-2c49-42ba-b353-721505b7d1e1" containerID="73fb1ec2d0e4d00edcefaae9cd750ae6ffa54046dcd3990daa3e62cbf311bb66" exitCode=0 Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.583370 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerDied","Data":"73fb1ec2d0e4d00edcefaae9cd750ae6ffa54046dcd3990daa3e62cbf311bb66"} Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.627400 4573 scope.go:117] "RemoveContainer" containerID="07e99f411f81ebd1c7d913a96bd39aa4350bc0201818a0cec99f71426fa35ab6" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.635638 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.667356 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.679331 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:37 crc kubenswrapper[4573]: E1203 09:04:37.679925 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-log" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.679950 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-log" Dec 03 09:04:37 crc kubenswrapper[4573]: E1203 09:04:37.679968 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-api" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.679978 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-api" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.680316 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-api" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.680355 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" containerName="nova-api-log" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.681760 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.687980 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.688216 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.688836 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.720118 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.771292 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.771379 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-config-data\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.771413 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-logs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.771456 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.771479 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smfnv\" (UniqueName: \"kubernetes.io/projected/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-kube-api-access-smfnv\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.771498 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.829789 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.873640 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.873772 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.873853 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-config-data\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.873887 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-logs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.873931 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.873951 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smfnv\" (UniqueName: \"kubernetes.io/projected/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-kube-api-access-smfnv\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.882625 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.882891 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-public-tls-certs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.885187 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-logs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.892139 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-config-data\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.892589 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smfnv\" (UniqueName: \"kubernetes.io/projected/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-kube-api-access-smfnv\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.898568 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " pod="openstack/nova-api-0" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.974877 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-log-httpd\") pod \"561f373f-2c49-42ba-b353-721505b7d1e1\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.975003 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-run-httpd\") pod \"561f373f-2c49-42ba-b353-721505b7d1e1\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.975040 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-ceilometer-tls-certs\") pod \"561f373f-2c49-42ba-b353-721505b7d1e1\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.975159 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-scripts\") pod \"561f373f-2c49-42ba-b353-721505b7d1e1\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.975231 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-sg-core-conf-yaml\") pod \"561f373f-2c49-42ba-b353-721505b7d1e1\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.975515 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "561f373f-2c49-42ba-b353-721505b7d1e1" (UID: "561f373f-2c49-42ba-b353-721505b7d1e1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.975565 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "561f373f-2c49-42ba-b353-721505b7d1e1" (UID: "561f373f-2c49-42ba-b353-721505b7d1e1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.975926 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-config-data\") pod \"561f373f-2c49-42ba-b353-721505b7d1e1\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.976003 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-combined-ca-bundle\") pod \"561f373f-2c49-42ba-b353-721505b7d1e1\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.976035 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvhx5\" (UniqueName: \"kubernetes.io/projected/561f373f-2c49-42ba-b353-721505b7d1e1-kube-api-access-pvhx5\") pod \"561f373f-2c49-42ba-b353-721505b7d1e1\" (UID: \"561f373f-2c49-42ba-b353-721505b7d1e1\") " Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.976547 4573 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.976567 4573 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561f373f-2c49-42ba-b353-721505b7d1e1-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.978941 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-scripts" (OuterVolumeSpecName: "scripts") pod "561f373f-2c49-42ba-b353-721505b7d1e1" (UID: "561f373f-2c49-42ba-b353-721505b7d1e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:37 crc kubenswrapper[4573]: I1203 09:04:37.992341 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561f373f-2c49-42ba-b353-721505b7d1e1-kube-api-access-pvhx5" (OuterVolumeSpecName: "kube-api-access-pvhx5") pod "561f373f-2c49-42ba-b353-721505b7d1e1" (UID: "561f373f-2c49-42ba-b353-721505b7d1e1"). InnerVolumeSpecName "kube-api-access-pvhx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.010075 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.023822 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "561f373f-2c49-42ba-b353-721505b7d1e1" (UID: "561f373f-2c49-42ba-b353-721505b7d1e1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.042856 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cda7757-bb11-46f2-9bac-a4f81a66bc1c" path="/var/lib/kubelet/pods/7cda7757-bb11-46f2-9bac-a4f81a66bc1c/volumes" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.066433 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "561f373f-2c49-42ba-b353-721505b7d1e1" (UID: "561f373f-2c49-42ba-b353-721505b7d1e1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.078803 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvhx5\" (UniqueName: \"kubernetes.io/projected/561f373f-2c49-42ba-b353-721505b7d1e1-kube-api-access-pvhx5\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.079014 4573 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.079120 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.079184 4573 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.112671 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "561f373f-2c49-42ba-b353-721505b7d1e1" (UID: "561f373f-2c49-42ba-b353-721505b7d1e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.166626 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-config-data" (OuterVolumeSpecName: "config-data") pod "561f373f-2c49-42ba-b353-721505b7d1e1" (UID: "561f373f-2c49-42ba-b353-721505b7d1e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.181495 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.181530 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561f373f-2c49-42ba-b353-721505b7d1e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.587879 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.597203 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561f373f-2c49-42ba-b353-721505b7d1e1","Type":"ContainerDied","Data":"801614fe2a78bd7c10bdf77dd84753cef3d83d72b76c88007292fc3a142a30a5"} Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.597270 4573 scope.go:117] "RemoveContainer" containerID="6aaf34b1cdeb041f67fc27c7f3a8f073f63ade7829322ee3e0adfe059ddea95a" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.597376 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: W1203 09:04:38.633618 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7cfc00c3_001a_4c8e_8cac_6e09158f88a5.slice/crio-e9cac9451f71ed269b11ff0a282f09ee4314c8c5d108a1eec712a28bb94f75fa WatchSource:0}: Error finding container e9cac9451f71ed269b11ff0a282f09ee4314c8c5d108a1eec712a28bb94f75fa: Status 404 returned error can't find the container with id e9cac9451f71ed269b11ff0a282f09ee4314c8c5d108a1eec712a28bb94f75fa Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.670856 4573 scope.go:117] "RemoveContainer" containerID="dc2588fd72227607c51a6e9d6ad483de1512389881979131ffb2a347fa23eb94" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.686825 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.706708 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.746797 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:38 crc kubenswrapper[4573]: E1203 09:04:38.747332 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="proxy-httpd" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.747352 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="proxy-httpd" Dec 03 09:04:38 crc kubenswrapper[4573]: E1203 09:04:38.747364 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="sg-core" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.747380 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="sg-core" Dec 03 09:04:38 crc kubenswrapper[4573]: E1203 09:04:38.747393 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="ceilometer-central-agent" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.747401 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="ceilometer-central-agent" Dec 03 09:04:38 crc kubenswrapper[4573]: E1203 09:04:38.747428 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="ceilometer-notification-agent" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.747434 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="ceilometer-notification-agent" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.747617 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="proxy-httpd" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.747628 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="ceilometer-central-agent" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.747645 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="ceilometer-notification-agent" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.747655 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" containerName="sg-core" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.749443 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.754009 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.754348 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.755983 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.769881 4573 scope.go:117] "RemoveContainer" containerID="73fb1ec2d0e4d00edcefaae9cd750ae6ffa54046dcd3990daa3e62cbf311bb66" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.774218 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.845017 4573 scope.go:117] "RemoveContainer" containerID="3899a4c5e236d6ea1934983729f4515e3cf4a60b9ac29e1075bd352c0c28590e" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.904559 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-config-data\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.904647 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3601284e-acb1-4e60-a4c1-18c404dc4072-log-httpd\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.904687 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf94f\" (UniqueName: \"kubernetes.io/projected/3601284e-acb1-4e60-a4c1-18c404dc4072-kube-api-access-jf94f\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.904738 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-scripts\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.904813 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.904834 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.904863 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3601284e-acb1-4e60-a4c1-18c404dc4072-run-httpd\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:38 crc kubenswrapper[4573]: I1203 09:04:38.904883 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.007307 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3601284e-acb1-4e60-a4c1-18c404dc4072-run-httpd\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.007650 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.007721 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-config-data\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.007772 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3601284e-acb1-4e60-a4c1-18c404dc4072-log-httpd\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.007805 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf94f\" (UniqueName: \"kubernetes.io/projected/3601284e-acb1-4e60-a4c1-18c404dc4072-kube-api-access-jf94f\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.007868 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-scripts\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.007981 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.008017 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.008217 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3601284e-acb1-4e60-a4c1-18c404dc4072-run-httpd\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.008692 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3601284e-acb1-4e60-a4c1-18c404dc4072-log-httpd\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.012438 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.013802 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.014516 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-config-data\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.022920 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-scripts\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.024884 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3601284e-acb1-4e60-a4c1-18c404dc4072-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.028521 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf94f\" (UniqueName: \"kubernetes.io/projected/3601284e-acb1-4e60-a4c1-18c404dc4072-kube-api-access-jf94f\") pod \"ceilometer-0\" (UID: \"3601284e-acb1-4e60-a4c1-18c404dc4072\") " pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.088221 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.581862 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 09:04:39 crc kubenswrapper[4573]: W1203 09:04:39.585558 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3601284e_acb1_4e60_a4c1_18c404dc4072.slice/crio-74e99d4905c2efb8d4edea9dce90d92f130321e5e1a4dd9d9e0c55bc288c353e WatchSource:0}: Error finding container 74e99d4905c2efb8d4edea9dce90d92f130321e5e1a4dd9d9e0c55bc288c353e: Status 404 returned error can't find the container with id 74e99d4905c2efb8d4edea9dce90d92f130321e5e1a4dd9d9e0c55bc288c353e Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.625807 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cfc00c3-001a-4c8e-8cac-6e09158f88a5","Type":"ContainerStarted","Data":"e80165f2b3e2b5b7c9860fdcc4edc5f7ec09b9385cd6e9050701b5398326b8e1"} Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.625854 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cfc00c3-001a-4c8e-8cac-6e09158f88a5","Type":"ContainerStarted","Data":"23854c9ecbf477398f42c2743c58be83f10d51284ddf3b7e209c1fd830a95282"} Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.625863 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cfc00c3-001a-4c8e-8cac-6e09158f88a5","Type":"ContainerStarted","Data":"e9cac9451f71ed269b11ff0a282f09ee4314c8c5d108a1eec712a28bb94f75fa"} Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.630595 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3601284e-acb1-4e60-a4c1-18c404dc4072","Type":"ContainerStarted","Data":"74e99d4905c2efb8d4edea9dce90d92f130321e5e1a4dd9d9e0c55bc288c353e"} Dec 03 09:04:39 crc kubenswrapper[4573]: I1203 09:04:39.650127 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.6501007039999998 podStartE2EDuration="2.650100704s" podCreationTimestamp="2025-12-03 09:04:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:39.642897706 +0000 UTC m=+1600.211276975" watchObservedRunningTime="2025-12-03 09:04:39.650100704 +0000 UTC m=+1600.218479963" Dec 03 09:04:40 crc kubenswrapper[4573]: I1203 09:04:40.041718 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561f373f-2c49-42ba-b353-721505b7d1e1" path="/var/lib/kubelet/pods/561f373f-2c49-42ba-b353-721505b7d1e1/volumes" Dec 03 09:04:40 crc kubenswrapper[4573]: I1203 09:04:40.209123 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:04:40 crc kubenswrapper[4573]: I1203 09:04:40.315374 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-lhxvs"] Dec 03 09:04:40 crc kubenswrapper[4573]: I1203 09:04:40.316469 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" podUID="8248677f-8e43-4fd1-92ed-560d826e6d60" containerName="dnsmasq-dns" containerID="cri-o://4b7e20ae8779c2c6711539056f922aafacf8665fd530121e91f4181b96ed33b5" gracePeriod=10 Dec 03 09:04:40 crc kubenswrapper[4573]: I1203 09:04:40.651714 4573 generic.go:334] "Generic (PLEG): container finished" podID="8248677f-8e43-4fd1-92ed-560d826e6d60" containerID="4b7e20ae8779c2c6711539056f922aafacf8665fd530121e91f4181b96ed33b5" exitCode=0 Dec 03 09:04:40 crc kubenswrapper[4573]: I1203 09:04:40.652143 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" event={"ID":"8248677f-8e43-4fd1-92ed-560d826e6d60","Type":"ContainerDied","Data":"4b7e20ae8779c2c6711539056f922aafacf8665fd530121e91f4181b96ed33b5"} Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.367530 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.494622 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-svc\") pod \"8248677f-8e43-4fd1-92ed-560d826e6d60\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.494688 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-swift-storage-0\") pod \"8248677f-8e43-4fd1-92ed-560d826e6d60\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.494774 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4srnd\" (UniqueName: \"kubernetes.io/projected/8248677f-8e43-4fd1-92ed-560d826e6d60-kube-api-access-4srnd\") pod \"8248677f-8e43-4fd1-92ed-560d826e6d60\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.494796 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-sb\") pod \"8248677f-8e43-4fd1-92ed-560d826e6d60\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.494881 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-nb\") pod \"8248677f-8e43-4fd1-92ed-560d826e6d60\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.494969 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-config\") pod \"8248677f-8e43-4fd1-92ed-560d826e6d60\" (UID: \"8248677f-8e43-4fd1-92ed-560d826e6d60\") " Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.501142 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8248677f-8e43-4fd1-92ed-560d826e6d60-kube-api-access-4srnd" (OuterVolumeSpecName: "kube-api-access-4srnd") pod "8248677f-8e43-4fd1-92ed-560d826e6d60" (UID: "8248677f-8e43-4fd1-92ed-560d826e6d60"). InnerVolumeSpecName "kube-api-access-4srnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.597341 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4srnd\" (UniqueName: \"kubernetes.io/projected/8248677f-8e43-4fd1-92ed-560d826e6d60-kube-api-access-4srnd\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.624671 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8248677f-8e43-4fd1-92ed-560d826e6d60" (UID: "8248677f-8e43-4fd1-92ed-560d826e6d60"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.629810 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8248677f-8e43-4fd1-92ed-560d826e6d60" (UID: "8248677f-8e43-4fd1-92ed-560d826e6d60"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.652162 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8248677f-8e43-4fd1-92ed-560d826e6d60" (UID: "8248677f-8e43-4fd1-92ed-560d826e6d60"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.657627 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-config" (OuterVolumeSpecName: "config") pod "8248677f-8e43-4fd1-92ed-560d826e6d60" (UID: "8248677f-8e43-4fd1-92ed-560d826e6d60"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.677156 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8248677f-8e43-4fd1-92ed-560d826e6d60" (UID: "8248677f-8e43-4fd1-92ed-560d826e6d60"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.689315 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3601284e-acb1-4e60-a4c1-18c404dc4072","Type":"ContainerStarted","Data":"1a6ccc81fa91db7ecd179d4df1f67741537c21210415fd335373f09938f98bc5"} Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.693231 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" event={"ID":"8248677f-8e43-4fd1-92ed-560d826e6d60","Type":"ContainerDied","Data":"732cd0a2e8caac2afcc63ff38227e64a34865c72d889adc88b088dc65d38455c"} Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.693284 4573 scope.go:117] "RemoveContainer" containerID="4b7e20ae8779c2c6711539056f922aafacf8665fd530121e91f4181b96ed33b5" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.693405 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-lhxvs" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.705427 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.705457 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.705467 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.705476 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.705486 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8248677f-8e43-4fd1-92ed-560d826e6d60-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.713924 4573 generic.go:334] "Generic (PLEG): container finished" podID="178dc07a-005e-4e3d-9c12-28b42eb2cc7d" containerID="04af37895750db5d16417f6b792e13afe02547877c30ce1dd17ba189134517bd" exitCode=0 Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.713977 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pgtv2" event={"ID":"178dc07a-005e-4e3d-9c12-28b42eb2cc7d","Type":"ContainerDied","Data":"04af37895750db5d16417f6b792e13afe02547877c30ce1dd17ba189134517bd"} Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.775769 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-lhxvs"] Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.780926 4573 scope.go:117] "RemoveContainer" containerID="18cdc7540bff79b8d406315a9abf428dea11eb97b58e3472946d4b079237bd3c" Dec 03 09:04:41 crc kubenswrapper[4573]: I1203 09:04:41.788910 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-lhxvs"] Dec 03 09:04:42 crc kubenswrapper[4573]: I1203 09:04:42.040945 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8248677f-8e43-4fd1-92ed-560d826e6d60" path="/var/lib/kubelet/pods/8248677f-8e43-4fd1-92ed-560d826e6d60/volumes" Dec 03 09:04:42 crc kubenswrapper[4573]: I1203 09:04:42.734209 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3601284e-acb1-4e60-a4c1-18c404dc4072","Type":"ContainerStarted","Data":"e557bfa6c4efe4b030de082bc92641f7d9b6d5f0ba3e3a594d61b5e892920005"} Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.144381 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.243489 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-scripts\") pod \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.243669 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59gpj\" (UniqueName: \"kubernetes.io/projected/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-kube-api-access-59gpj\") pod \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.243860 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-combined-ca-bundle\") pod \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.243891 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-config-data\") pod \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\" (UID: \"178dc07a-005e-4e3d-9c12-28b42eb2cc7d\") " Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.251037 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-kube-api-access-59gpj" (OuterVolumeSpecName: "kube-api-access-59gpj") pod "178dc07a-005e-4e3d-9c12-28b42eb2cc7d" (UID: "178dc07a-005e-4e3d-9c12-28b42eb2cc7d"). InnerVolumeSpecName "kube-api-access-59gpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.251168 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-scripts" (OuterVolumeSpecName: "scripts") pod "178dc07a-005e-4e3d-9c12-28b42eb2cc7d" (UID: "178dc07a-005e-4e3d-9c12-28b42eb2cc7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.288891 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-config-data" (OuterVolumeSpecName: "config-data") pod "178dc07a-005e-4e3d-9c12-28b42eb2cc7d" (UID: "178dc07a-005e-4e3d-9c12-28b42eb2cc7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.295903 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "178dc07a-005e-4e3d-9c12-28b42eb2cc7d" (UID: "178dc07a-005e-4e3d-9c12-28b42eb2cc7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.346287 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.346355 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.346365 4573 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.346375 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59gpj\" (UniqueName: \"kubernetes.io/projected/178dc07a-005e-4e3d-9c12-28b42eb2cc7d-kube-api-access-59gpj\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.710606 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.766762 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-pgtv2" event={"ID":"178dc07a-005e-4e3d-9c12-28b42eb2cc7d","Type":"ContainerDied","Data":"baf6a43b27f8c671b6eee22655f1fc9365ec2849fd04a9dd3db244e88cdc18bc"} Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.766804 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baf6a43b27f8c671b6eee22655f1fc9365ec2849fd04a9dd3db244e88cdc18bc" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.766865 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-pgtv2" Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.810762 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3601284e-acb1-4e60-a4c1-18c404dc4072","Type":"ContainerStarted","Data":"5bd4e17f0f7d69e83f1073fec62aa9390671061b3c19fdd3d3896ab8b4ac8ea8"} Dec 03 09:04:43 crc kubenswrapper[4573]: I1203 09:04:43.832673 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tmzcx" Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.000250 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.001127 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerName="nova-api-log" containerID="cri-o://23854c9ecbf477398f42c2743c58be83f10d51284ddf3b7e209c1fd830a95282" gracePeriod=30 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.001572 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerName="nova-api-api" containerID="cri-o://e80165f2b3e2b5b7c9860fdcc4edc5f7ec09b9385cd6e9050701b5398326b8e1" gracePeriod=30 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.092015 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.092238 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="70534db4-de8c-4c04-92d6-60abef7c4843" containerName="nova-scheduler-scheduler" containerID="cri-o://9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" gracePeriod=30 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.115620 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.116593 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-log" containerID="cri-o://5b179ffe542f3b6ae273d19064c9aa5e5c205e88e6a09449ab437659c8fb9c00" gracePeriod=30 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.117077 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-metadata" containerID="cri-o://5d9d61a43c20e0d52583c62fc4d630055a60204687c14c1f2855cd493ef8b399" gracePeriod=30 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.145812 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tmzcx"] Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.382837 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tcg9b"] Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.386999 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tcg9b" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="registry-server" containerID="cri-o://c409c85639c839576ef225c3b40d7c68e9824fe4ebf7d8c064837b0bbe4a17b7" gracePeriod=2 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.842535 4573 generic.go:334] "Generic (PLEG): container finished" podID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerID="c409c85639c839576ef225c3b40d7c68e9824fe4ebf7d8c064837b0bbe4a17b7" exitCode=0 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.844938 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tcg9b" event={"ID":"09312f5f-d77a-431f-b16b-f31a9cde4335","Type":"ContainerDied","Data":"c409c85639c839576ef225c3b40d7c68e9824fe4ebf7d8c064837b0bbe4a17b7"} Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.849101 4573 generic.go:334] "Generic (PLEG): container finished" podID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerID="23854c9ecbf477398f42c2743c58be83f10d51284ddf3b7e209c1fd830a95282" exitCode=143 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.849163 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cfc00c3-001a-4c8e-8cac-6e09158f88a5","Type":"ContainerDied","Data":"23854c9ecbf477398f42c2743c58be83f10d51284ddf3b7e209c1fd830a95282"} Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.858458 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3601284e-acb1-4e60-a4c1-18c404dc4072","Type":"ContainerStarted","Data":"ab6577057203b0364d1a5c400382e25c77ed6d41b46907d2c688f7ddbfc56508"} Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.860542 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.883943 4573 generic.go:334] "Generic (PLEG): container finished" podID="4e482326-82c9-4621-896b-dcd607248b36" containerID="5b179ffe542f3b6ae273d19064c9aa5e5c205e88e6a09449ab437659c8fb9c00" exitCode=143 Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.885289 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e482326-82c9-4621-896b-dcd607248b36","Type":"ContainerDied","Data":"5b179ffe542f3b6ae273d19064c9aa5e5c205e88e6a09449ab437659c8fb9c00"} Dec 03 09:04:44 crc kubenswrapper[4573]: I1203 09:04:44.901183 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.177530581 podStartE2EDuration="6.901163203s" podCreationTimestamp="2025-12-03 09:04:38 +0000 UTC" firstStartedPulling="2025-12-03 09:04:39.588599934 +0000 UTC m=+1600.156979183" lastFinishedPulling="2025-12-03 09:04:44.312232546 +0000 UTC m=+1604.880611805" observedRunningTime="2025-12-03 09:04:44.895785226 +0000 UTC m=+1605.464164485" watchObservedRunningTime="2025-12-03 09:04:44.901163203 +0000 UTC m=+1605.469542462" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.053550 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.223740 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr2gv\" (UniqueName: \"kubernetes.io/projected/09312f5f-d77a-431f-b16b-f31a9cde4335-kube-api-access-kr2gv\") pod \"09312f5f-d77a-431f-b16b-f31a9cde4335\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.223978 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-catalog-content\") pod \"09312f5f-d77a-431f-b16b-f31a9cde4335\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.224004 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-utilities\") pod \"09312f5f-d77a-431f-b16b-f31a9cde4335\" (UID: \"09312f5f-d77a-431f-b16b-f31a9cde4335\") " Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.227104 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-utilities" (OuterVolumeSpecName: "utilities") pod "09312f5f-d77a-431f-b16b-f31a9cde4335" (UID: "09312f5f-d77a-431f-b16b-f31a9cde4335"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.233778 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09312f5f-d77a-431f-b16b-f31a9cde4335-kube-api-access-kr2gv" (OuterVolumeSpecName: "kube-api-access-kr2gv") pod "09312f5f-d77a-431f-b16b-f31a9cde4335" (UID: "09312f5f-d77a-431f-b16b-f31a9cde4335"). InnerVolumeSpecName "kube-api-access-kr2gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.326744 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.326773 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr2gv\" (UniqueName: \"kubernetes.io/projected/09312f5f-d77a-431f-b16b-f31a9cde4335-kube-api-access-kr2gv\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.452728 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09312f5f-d77a-431f-b16b-f31a9cde4335" (UID: "09312f5f-d77a-431f-b16b-f31a9cde4335"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.542753 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09312f5f-d77a-431f-b16b-f31a9cde4335-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.918687 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tcg9b" event={"ID":"09312f5f-d77a-431f-b16b-f31a9cde4335","Type":"ContainerDied","Data":"4282c6de181ce0b9211c01e994b3328d7b9742d36706666222f08050b835f1ff"} Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.919011 4573 scope.go:117] "RemoveContainer" containerID="c409c85639c839576ef225c3b40d7c68e9824fe4ebf7d8c064837b0bbe4a17b7" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.919303 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tcg9b" Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.936213 4573 generic.go:334] "Generic (PLEG): container finished" podID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerID="e80165f2b3e2b5b7c9860fdcc4edc5f7ec09b9385cd6e9050701b5398326b8e1" exitCode=0 Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.936838 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cfc00c3-001a-4c8e-8cac-6e09158f88a5","Type":"ContainerDied","Data":"e80165f2b3e2b5b7c9860fdcc4edc5f7ec09b9385cd6e9050701b5398326b8e1"} Dec 03 09:04:45 crc kubenswrapper[4573]: I1203 09:04:45.994122 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tcg9b"] Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.014243 4573 scope.go:117] "RemoveContainer" containerID="aaf385ffabe1b2acf8d5cce355ea2c97b0b2d07a0d2e2a31a2fb2124399c8bcb" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.015234 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tcg9b"] Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.062173 4573 scope.go:117] "RemoveContainer" containerID="b38d21b10f8431cfb8eb6c66a01924317e1c93f07db424d53d81d5380932e085" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.080479 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" path="/var/lib/kubelet/pods/09312f5f-d77a-431f-b16b-f31a9cde4335/volumes" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.218511 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.269431 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smfnv\" (UniqueName: \"kubernetes.io/projected/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-kube-api-access-smfnv\") pod \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.269603 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-public-tls-certs\") pod \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.269688 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-logs\") pod \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.269765 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-combined-ca-bundle\") pod \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.269821 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-internal-tls-certs\") pod \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.269856 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-config-data\") pod \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\" (UID: \"7cfc00c3-001a-4c8e-8cac-6e09158f88a5\") " Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.270096 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-logs" (OuterVolumeSpecName: "logs") pod "7cfc00c3-001a-4c8e-8cac-6e09158f88a5" (UID: "7cfc00c3-001a-4c8e-8cac-6e09158f88a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.270431 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.284850 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-kube-api-access-smfnv" (OuterVolumeSpecName: "kube-api-access-smfnv") pod "7cfc00c3-001a-4c8e-8cac-6e09158f88a5" (UID: "7cfc00c3-001a-4c8e-8cac-6e09158f88a5"). InnerVolumeSpecName "kube-api-access-smfnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.322386 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cfc00c3-001a-4c8e-8cac-6e09158f88a5" (UID: "7cfc00c3-001a-4c8e-8cac-6e09158f88a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.366212 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-config-data" (OuterVolumeSpecName: "config-data") pod "7cfc00c3-001a-4c8e-8cac-6e09158f88a5" (UID: "7cfc00c3-001a-4c8e-8cac-6e09158f88a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.371933 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.371961 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.371973 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smfnv\" (UniqueName: \"kubernetes.io/projected/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-kube-api-access-smfnv\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.426075 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7cfc00c3-001a-4c8e-8cac-6e09158f88a5" (UID: "7cfc00c3-001a-4c8e-8cac-6e09158f88a5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.435015 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7cfc00c3-001a-4c8e-8cac-6e09158f88a5" (UID: "7cfc00c3-001a-4c8e-8cac-6e09158f88a5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.474145 4573 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.474180 4573 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cfc00c3-001a-4c8e-8cac-6e09158f88a5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:46 crc kubenswrapper[4573]: E1203 09:04:46.596146 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:46 crc kubenswrapper[4573]: E1203 09:04:46.598765 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:46 crc kubenswrapper[4573]: E1203 09:04:46.600603 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:46 crc kubenswrapper[4573]: E1203 09:04:46.600638 4573 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="70534db4-de8c-4c04-92d6-60abef7c4843" containerName="nova-scheduler-scheduler" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.948703 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7cfc00c3-001a-4c8e-8cac-6e09158f88a5","Type":"ContainerDied","Data":"e9cac9451f71ed269b11ff0a282f09ee4314c8c5d108a1eec712a28bb94f75fa"} Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.949015 4573 scope.go:117] "RemoveContainer" containerID="e80165f2b3e2b5b7c9860fdcc4edc5f7ec09b9385cd6e9050701b5398326b8e1" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.948720 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:46 crc kubenswrapper[4573]: I1203 09:04:46.995700 4573 scope.go:117] "RemoveContainer" containerID="23854c9ecbf477398f42c2743c58be83f10d51284ddf3b7e209c1fd830a95282" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.012628 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.034126 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.052928 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:47 crc kubenswrapper[4573]: E1203 09:04:47.054282 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="extract-utilities" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054308 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="extract-utilities" Dec 03 09:04:47 crc kubenswrapper[4573]: E1203 09:04:47.054324 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8248677f-8e43-4fd1-92ed-560d826e6d60" containerName="init" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054330 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8248677f-8e43-4fd1-92ed-560d826e6d60" containerName="init" Dec 03 09:04:47 crc kubenswrapper[4573]: E1203 09:04:47.054337 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerName="nova-api-log" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054343 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerName="nova-api-log" Dec 03 09:04:47 crc kubenswrapper[4573]: E1203 09:04:47.054360 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8248677f-8e43-4fd1-92ed-560d826e6d60" containerName="dnsmasq-dns" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054368 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8248677f-8e43-4fd1-92ed-560d826e6d60" containerName="dnsmasq-dns" Dec 03 09:04:47 crc kubenswrapper[4573]: E1203 09:04:47.054389 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="registry-server" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054395 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="registry-server" Dec 03 09:04:47 crc kubenswrapper[4573]: E1203 09:04:47.054407 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178dc07a-005e-4e3d-9c12-28b42eb2cc7d" containerName="nova-manage" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054413 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="178dc07a-005e-4e3d-9c12-28b42eb2cc7d" containerName="nova-manage" Dec 03 09:04:47 crc kubenswrapper[4573]: E1203 09:04:47.054426 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="extract-content" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054432 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="extract-content" Dec 03 09:04:47 crc kubenswrapper[4573]: E1203 09:04:47.054445 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerName="nova-api-api" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054450 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerName="nova-api-api" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054621 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerName="nova-api-api" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054637 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8248677f-8e43-4fd1-92ed-560d826e6d60" containerName="dnsmasq-dns" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054656 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="178dc07a-005e-4e3d-9c12-28b42eb2cc7d" containerName="nova-manage" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054670 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" containerName="nova-api-log" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.054680 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="09312f5f-d77a-431f-b16b-f31a9cde4335" containerName="registry-server" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.055798 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.061986 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.062348 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.062467 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.083386 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.092155 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b06177f1-1c12-41d9-b7f8-7df5e6986872-logs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.092212 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.092257 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-public-tls-certs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.092282 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcqvz\" (UniqueName: \"kubernetes.io/projected/b06177f1-1c12-41d9-b7f8-7df5e6986872-kube-api-access-mcqvz\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.092347 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-config-data\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.092433 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.197543 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-public-tls-certs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.197622 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcqvz\" (UniqueName: \"kubernetes.io/projected/b06177f1-1c12-41d9-b7f8-7df5e6986872-kube-api-access-mcqvz\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.198099 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-config-data\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.198256 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.198402 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b06177f1-1c12-41d9-b7f8-7df5e6986872-logs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.198454 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.199560 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b06177f1-1c12-41d9-b7f8-7df5e6986872-logs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.205707 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-config-data\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.211839 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.217812 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.222412 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b06177f1-1c12-41d9-b7f8-7df5e6986872-public-tls-certs\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.228456 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcqvz\" (UniqueName: \"kubernetes.io/projected/b06177f1-1c12-41d9-b7f8-7df5e6986872-kube-api-access-mcqvz\") pod \"nova-api-0\" (UID: \"b06177f1-1c12-41d9-b7f8-7df5e6986872\") " pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.391656 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.640600 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:51388->10.217.0.199:8775: read: connection reset by peer" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.641186 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:51392->10.217.0.199:8775: read: connection reset by peer" Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.964868 4573 generic.go:334] "Generic (PLEG): container finished" podID="4e482326-82c9-4621-896b-dcd607248b36" containerID="5d9d61a43c20e0d52583c62fc4d630055a60204687c14c1f2855cd493ef8b399" exitCode=0 Dec 03 09:04:47 crc kubenswrapper[4573]: I1203 09:04:47.965209 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e482326-82c9-4621-896b-dcd607248b36","Type":"ContainerDied","Data":"5d9d61a43c20e0d52583c62fc4d630055a60204687c14c1f2855cd493ef8b399"} Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.081000 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cfc00c3-001a-4c8e-8cac-6e09158f88a5" path="/var/lib/kubelet/pods/7cfc00c3-001a-4c8e-8cac-6e09158f88a5/volumes" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.161689 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.382226 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.436472 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-config-data\") pod \"4e482326-82c9-4621-896b-dcd607248b36\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.506497 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-config-data" (OuterVolumeSpecName: "config-data") pod "4e482326-82c9-4621-896b-dcd607248b36" (UID: "4e482326-82c9-4621-896b-dcd607248b36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.547064 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-nova-metadata-tls-certs\") pod \"4e482326-82c9-4621-896b-dcd607248b36\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.547124 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e482326-82c9-4621-896b-dcd607248b36-logs\") pod \"4e482326-82c9-4621-896b-dcd607248b36\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.547150 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghnmn\" (UniqueName: \"kubernetes.io/projected/4e482326-82c9-4621-896b-dcd607248b36-kube-api-access-ghnmn\") pod \"4e482326-82c9-4621-896b-dcd607248b36\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.547489 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-combined-ca-bundle\") pod \"4e482326-82c9-4621-896b-dcd607248b36\" (UID: \"4e482326-82c9-4621-896b-dcd607248b36\") " Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.547896 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.548110 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e482326-82c9-4621-896b-dcd607248b36-logs" (OuterVolumeSpecName: "logs") pod "4e482326-82c9-4621-896b-dcd607248b36" (UID: "4e482326-82c9-4621-896b-dcd607248b36"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.575297 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e482326-82c9-4621-896b-dcd607248b36-kube-api-access-ghnmn" (OuterVolumeSpecName: "kube-api-access-ghnmn") pod "4e482326-82c9-4621-896b-dcd607248b36" (UID: "4e482326-82c9-4621-896b-dcd607248b36"). InnerVolumeSpecName "kube-api-access-ghnmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.603452 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e482326-82c9-4621-896b-dcd607248b36" (UID: "4e482326-82c9-4621-896b-dcd607248b36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.650518 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.650562 4573 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e482326-82c9-4621-896b-dcd607248b36-logs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.650573 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghnmn\" (UniqueName: \"kubernetes.io/projected/4e482326-82c9-4621-896b-dcd607248b36-kube-api-access-ghnmn\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.688182 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4e482326-82c9-4621-896b-dcd607248b36" (UID: "4e482326-82c9-4621-896b-dcd607248b36"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.753250 4573 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e482326-82c9-4621-896b-dcd607248b36-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.986336 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4e482326-82c9-4621-896b-dcd607248b36","Type":"ContainerDied","Data":"13e1413baf1abfa15486d0b6ae9e55dbe409b81857fbeed7ac00af9f13a63eb9"} Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.986704 4573 scope.go:117] "RemoveContainer" containerID="5d9d61a43c20e0d52583c62fc4d630055a60204687c14c1f2855cd493ef8b399" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.986628 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.994841 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b06177f1-1c12-41d9-b7f8-7df5e6986872","Type":"ContainerStarted","Data":"e41b86d92ce99b4ee8ba972daea09abcf5f9ff41d1636170e9004a78606b28a4"} Dec 03 09:04:48 crc kubenswrapper[4573]: I1203 09:04:48.994897 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b06177f1-1c12-41d9-b7f8-7df5e6986872","Type":"ContainerStarted","Data":"5a61ff4319d36103420240b2f1d9bc989cc2d1140c3f1b7b8d702afb4b6fe6a8"} Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.018785 4573 scope.go:117] "RemoveContainer" containerID="5b179ffe542f3b6ae273d19064c9aa5e5c205e88e6a09449ab437659c8fb9c00" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.040557 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.063343 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.081461 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:49 crc kubenswrapper[4573]: E1203 09:04:49.082039 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-metadata" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.082088 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-metadata" Dec 03 09:04:49 crc kubenswrapper[4573]: E1203 09:04:49.082109 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-log" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.082118 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-log" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.082415 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-log" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.082465 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e482326-82c9-4621-896b-dcd607248b36" containerName="nova-metadata-metadata" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.083847 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.087034 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.089118 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.110606 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.164591 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js2ng\" (UniqueName: \"kubernetes.io/projected/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-kube-api-access-js2ng\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.164658 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-config-data\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.164708 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.164726 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-logs\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.164775 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.266995 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js2ng\" (UniqueName: \"kubernetes.io/projected/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-kube-api-access-js2ng\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.267123 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-config-data\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.267190 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.267216 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-logs\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.267291 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.268143 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-logs\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.273998 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-config-data\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.275905 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.280890 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.287258 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js2ng\" (UniqueName: \"kubernetes.io/projected/f233cf9b-b14d-4eee-8f84-3fcb5d85f571-kube-api-access-js2ng\") pod \"nova-metadata-0\" (UID: \"f233cf9b-b14d-4eee-8f84-3fcb5d85f571\") " pod="openstack/nova-metadata-0" Dec 03 09:04:49 crc kubenswrapper[4573]: I1203 09:04:49.416816 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 09:04:50 crc kubenswrapper[4573]: I1203 09:04:50.076142 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e482326-82c9-4621-896b-dcd607248b36" path="/var/lib/kubelet/pods/4e482326-82c9-4621-896b-dcd607248b36/volumes" Dec 03 09:04:50 crc kubenswrapper[4573]: I1203 09:04:50.077015 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b06177f1-1c12-41d9-b7f8-7df5e6986872","Type":"ContainerStarted","Data":"d9cb93a4ad51b434cc185bfeeef8a18b3ab8b406dc57fba301a643c830712080"} Dec 03 09:04:50 crc kubenswrapper[4573]: I1203 09:04:50.168916 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 09:04:50 crc kubenswrapper[4573]: I1203 09:04:50.182343 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.1823151 podStartE2EDuration="4.1823151s" podCreationTimestamp="2025-12-03 09:04:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:50.156337295 +0000 UTC m=+1610.724716554" watchObservedRunningTime="2025-12-03 09:04:50.1823151 +0000 UTC m=+1610.750694359" Dec 03 09:04:50 crc kubenswrapper[4573]: I1203 09:04:50.731578 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96" containerName="galera" probeResult="failure" output="command timed out" Dec 03 09:04:50 crc kubenswrapper[4573]: I1203 09:04:50.733981 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96" containerName="galera" probeResult="failure" output="command timed out" Dec 03 09:04:51 crc kubenswrapper[4573]: I1203 09:04:51.066178 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f233cf9b-b14d-4eee-8f84-3fcb5d85f571","Type":"ContainerStarted","Data":"40144b8e634159a58d2e6201338aea5f4ae4887cffaa11ac46477d3671ebc54e"} Dec 03 09:04:51 crc kubenswrapper[4573]: I1203 09:04:51.066232 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f233cf9b-b14d-4eee-8f84-3fcb5d85f571","Type":"ContainerStarted","Data":"37f27908a74db64cb9447cefc6c314b2615943043a88e693a2dde2ebe7a81123"} Dec 03 09:04:51 crc kubenswrapper[4573]: I1203 09:04:51.066247 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f233cf9b-b14d-4eee-8f84-3fcb5d85f571","Type":"ContainerStarted","Data":"a8008d15ddef00352097dc33352b039d2eb08c547a5fe8169546ebc51cfa295c"} Dec 03 09:04:51 crc kubenswrapper[4573]: E1203 09:04:51.598300 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535 is running failed: container process not found" containerID="9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:51 crc kubenswrapper[4573]: E1203 09:04:51.601785 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535 is running failed: container process not found" containerID="9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:51 crc kubenswrapper[4573]: E1203 09:04:51.602492 4573 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535 is running failed: container process not found" containerID="9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 09:04:51 crc kubenswrapper[4573]: E1203 09:04:51.602539 4573 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="70534db4-de8c-4c04-92d6-60abef7c4843" containerName="nova-scheduler-scheduler" Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.134618 4573 generic.go:334] "Generic (PLEG): container finished" podID="70534db4-de8c-4c04-92d6-60abef7c4843" containerID="9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" exitCode=0 Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.135596 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70534db4-de8c-4c04-92d6-60abef7c4843","Type":"ContainerDied","Data":"9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535"} Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.271569 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.329953 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.329913138 podStartE2EDuration="3.329913138s" podCreationTimestamp="2025-12-03 09:04:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:51.099038677 +0000 UTC m=+1611.667417956" watchObservedRunningTime="2025-12-03 09:04:52.329913138 +0000 UTC m=+1612.898292397" Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.376117 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-combined-ca-bundle\") pod \"70534db4-de8c-4c04-92d6-60abef7c4843\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.376266 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5x7f\" (UniqueName: \"kubernetes.io/projected/70534db4-de8c-4c04-92d6-60abef7c4843-kube-api-access-s5x7f\") pod \"70534db4-de8c-4c04-92d6-60abef7c4843\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.376340 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-config-data\") pod \"70534db4-de8c-4c04-92d6-60abef7c4843\" (UID: \"70534db4-de8c-4c04-92d6-60abef7c4843\") " Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.430493 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70534db4-de8c-4c04-92d6-60abef7c4843-kube-api-access-s5x7f" (OuterVolumeSpecName: "kube-api-access-s5x7f") pod "70534db4-de8c-4c04-92d6-60abef7c4843" (UID: "70534db4-de8c-4c04-92d6-60abef7c4843"). InnerVolumeSpecName "kube-api-access-s5x7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.475560 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-config-data" (OuterVolumeSpecName: "config-data") pod "70534db4-de8c-4c04-92d6-60abef7c4843" (UID: "70534db4-de8c-4c04-92d6-60abef7c4843"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.476259 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70534db4-de8c-4c04-92d6-60abef7c4843" (UID: "70534db4-de8c-4c04-92d6-60abef7c4843"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.479100 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5x7f\" (UniqueName: \"kubernetes.io/projected/70534db4-de8c-4c04-92d6-60abef7c4843-kube-api-access-s5x7f\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.479144 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:52 crc kubenswrapper[4573]: I1203 09:04:52.479159 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70534db4-de8c-4c04-92d6-60abef7c4843-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.144687 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70534db4-de8c-4c04-92d6-60abef7c4843","Type":"ContainerDied","Data":"c4900c7ee1ca64fa9598cf359330da86f3079491c4f423096efad4616a4f69b5"} Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.144765 4573 scope.go:117] "RemoveContainer" containerID="9e58eba585a56abc91e028608dc99420518fc6257751a927012b183c0bffb535" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.144790 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.184027 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.196098 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.217075 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:53 crc kubenswrapper[4573]: E1203 09:04:53.217644 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70534db4-de8c-4c04-92d6-60abef7c4843" containerName="nova-scheduler-scheduler" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.217669 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="70534db4-de8c-4c04-92d6-60abef7c4843" containerName="nova-scheduler-scheduler" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.217901 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="70534db4-de8c-4c04-92d6-60abef7c4843" containerName="nova-scheduler-scheduler" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.219450 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.223740 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.248268 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.398296 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f569d8-50e2-4177-bec8-8225de477357-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.398396 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f569d8-50e2-4177-bec8-8225de477357-config-data\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.398464 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h2kx\" (UniqueName: \"kubernetes.io/projected/56f569d8-50e2-4177-bec8-8225de477357-kube-api-access-4h2kx\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.500521 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f569d8-50e2-4177-bec8-8225de477357-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.501496 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f569d8-50e2-4177-bec8-8225de477357-config-data\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.501737 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h2kx\" (UniqueName: \"kubernetes.io/projected/56f569d8-50e2-4177-bec8-8225de477357-kube-api-access-4h2kx\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.510437 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f569d8-50e2-4177-bec8-8225de477357-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.512729 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f569d8-50e2-4177-bec8-8225de477357-config-data\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.533706 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h2kx\" (UniqueName: \"kubernetes.io/projected/56f569d8-50e2-4177-bec8-8225de477357-kube-api-access-4h2kx\") pod \"nova-scheduler-0\" (UID: \"56f569d8-50e2-4177-bec8-8225de477357\") " pod="openstack/nova-scheduler-0" Dec 03 09:04:53 crc kubenswrapper[4573]: I1203 09:04:53.536536 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 09:04:54 crc kubenswrapper[4573]: I1203 09:04:54.044557 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70534db4-de8c-4c04-92d6-60abef7c4843" path="/var/lib/kubelet/pods/70534db4-de8c-4c04-92d6-60abef7c4843/volumes" Dec 03 09:04:54 crc kubenswrapper[4573]: I1203 09:04:54.092017 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 09:04:54 crc kubenswrapper[4573]: W1203 09:04:54.093361 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56f569d8_50e2_4177_bec8_8225de477357.slice/crio-9f937352a0afe18e36b64b717c4993856ce0db0b08da16a4f6e34f2dd96bccb5 WatchSource:0}: Error finding container 9f937352a0afe18e36b64b717c4993856ce0db0b08da16a4f6e34f2dd96bccb5: Status 404 returned error can't find the container with id 9f937352a0afe18e36b64b717c4993856ce0db0b08da16a4f6e34f2dd96bccb5 Dec 03 09:04:54 crc kubenswrapper[4573]: I1203 09:04:54.162547 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"56f569d8-50e2-4177-bec8-8225de477357","Type":"ContainerStarted","Data":"9f937352a0afe18e36b64b717c4993856ce0db0b08da16a4f6e34f2dd96bccb5"} Dec 03 09:04:54 crc kubenswrapper[4573]: I1203 09:04:54.417682 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 09:04:54 crc kubenswrapper[4573]: I1203 09:04:54.418132 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 09:04:55 crc kubenswrapper[4573]: I1203 09:04:55.177260 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"56f569d8-50e2-4177-bec8-8225de477357","Type":"ContainerStarted","Data":"0d34f6078d31fdf349260261122be9fb849d5e21a3a37442e41cd54955caeaa5"} Dec 03 09:04:56 crc kubenswrapper[4573]: I1203 09:04:56.944027 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:04:56 crc kubenswrapper[4573]: I1203 09:04:56.944452 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:04:57 crc kubenswrapper[4573]: I1203 09:04:57.392782 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 09:04:57 crc kubenswrapper[4573]: I1203 09:04:57.392838 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 09:04:58 crc kubenswrapper[4573]: I1203 09:04:58.406222 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b06177f1-1c12-41d9-b7f8-7df5e6986872" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:04:58 crc kubenswrapper[4573]: I1203 09:04:58.406268 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b06177f1-1c12-41d9-b7f8-7df5e6986872" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:04:58 crc kubenswrapper[4573]: I1203 09:04:58.537936 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 09:04:59 crc kubenswrapper[4573]: I1203 09:04:59.418212 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 09:04:59 crc kubenswrapper[4573]: I1203 09:04:59.418266 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 09:05:00 crc kubenswrapper[4573]: I1203 09:05:00.431280 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f233cf9b-b14d-4eee-8f84-3fcb5d85f571" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:05:00 crc kubenswrapper[4573]: I1203 09:05:00.432247 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f233cf9b-b14d-4eee-8f84-3fcb5d85f571" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:05:03 crc kubenswrapper[4573]: I1203 09:05:03.538089 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 09:05:03 crc kubenswrapper[4573]: I1203 09:05:03.570306 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 09:05:03 crc kubenswrapper[4573]: I1203 09:05:03.592869 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=10.592846919 podStartE2EDuration="10.592846919s" podCreationTimestamp="2025-12-03 09:04:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:04:55.200875489 +0000 UTC m=+1615.769254748" watchObservedRunningTime="2025-12-03 09:05:03.592846919 +0000 UTC m=+1624.161226178" Dec 03 09:05:04 crc kubenswrapper[4573]: I1203 09:05:04.258288 4573 scope.go:117] "RemoveContainer" containerID="ed5f5a3c9f995f9aebe30e238d48f500aa8378deb069dd20fc1d4a112f219619" Dec 03 09:05:04 crc kubenswrapper[4573]: I1203 09:05:04.288142 4573 scope.go:117] "RemoveContainer" containerID="e6f06726d144ccb117ad2ea1372333c2a0959919e71f9719caa0435ec7809e20" Dec 03 09:05:04 crc kubenswrapper[4573]: I1203 09:05:04.315618 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 09:05:04 crc kubenswrapper[4573]: E1203 09:05:04.948769 4573 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 03 09:05:07 crc kubenswrapper[4573]: I1203 09:05:07.403966 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 09:05:07 crc kubenswrapper[4573]: I1203 09:05:07.405877 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 09:05:07 crc kubenswrapper[4573]: I1203 09:05:07.406100 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 09:05:07 crc kubenswrapper[4573]: I1203 09:05:07.415204 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 09:05:08 crc kubenswrapper[4573]: I1203 09:05:08.317982 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 09:05:08 crc kubenswrapper[4573]: I1203 09:05:08.327692 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 09:05:09 crc kubenswrapper[4573]: I1203 09:05:09.370213 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 09:05:09 crc kubenswrapper[4573]: I1203 09:05:09.441357 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 09:05:09 crc kubenswrapper[4573]: I1203 09:05:09.442563 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 09:05:09 crc kubenswrapper[4573]: I1203 09:05:09.454748 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 09:05:10 crc kubenswrapper[4573]: I1203 09:05:10.354322 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 09:05:19 crc kubenswrapper[4573]: I1203 09:05:19.877373 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 09:05:21 crc kubenswrapper[4573]: I1203 09:05:21.138642 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 09:05:25 crc kubenswrapper[4573]: I1203 09:05:25.842244 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerName="rabbitmq" containerID="cri-o://8fc49a0e39d5014a0145f26bd4d0be966ee9fad4deaaf9c035afcb09e6c961b0" gracePeriod=604795 Dec 03 09:05:26 crc kubenswrapper[4573]: I1203 09:05:26.377717 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerName="rabbitmq" containerID="cri-o://5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64" gracePeriod=604795 Dec 03 09:05:26 crc kubenswrapper[4573]: I1203 09:05:26.943473 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:05:26 crc kubenswrapper[4573]: I1203 09:05:26.943559 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:05:28 crc kubenswrapper[4573]: I1203 09:05:28.162005 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 03 09:05:28 crc kubenswrapper[4573]: I1203 09:05:28.794334 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.551287 4573 generic.go:334] "Generic (PLEG): container finished" podID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerID="8fc49a0e39d5014a0145f26bd4d0be966ee9fad4deaaf9c035afcb09e6c961b0" exitCode=0 Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.552597 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7087e292-a9fd-4420-bb4a-eb2d50355082","Type":"ContainerDied","Data":"8fc49a0e39d5014a0145f26bd4d0be966ee9fad4deaaf9c035afcb09e6c961b0"} Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.552677 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7087e292-a9fd-4420-bb4a-eb2d50355082","Type":"ContainerDied","Data":"96c92813c3c463f20bb4b59d7c2ca49ce9e599f1e78b3e0f5e8d44723cca1e02"} Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.552733 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96c92813c3c463f20bb4b59d7c2ca49ce9e599f1e78b3e0f5e8d44723cca1e02" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.666956 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.835854 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7087e292-a9fd-4420-bb4a-eb2d50355082-pod-info\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836217 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-plugins-conf\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836271 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-confd\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836326 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7087e292-a9fd-4420-bb4a-eb2d50355082-erlang-cookie-secret\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836409 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-tls\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836428 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-config-data\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836512 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-plugins\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836530 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-server-conf\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836628 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-erlang-cookie\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836658 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kzkl\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-kube-api-access-5kzkl\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.836699 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7087e292-a9fd-4420-bb4a-eb2d50355082\" (UID: \"7087e292-a9fd-4420-bb4a-eb2d50355082\") " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.840236 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.854536 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.855370 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.862892 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.863342 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7087e292-a9fd-4420-bb4a-eb2d50355082-pod-info" (OuterVolumeSpecName: "pod-info") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.875120 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.896992 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-kube-api-access-5kzkl" (OuterVolumeSpecName: "kube-api-access-5kzkl") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "kube-api-access-5kzkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.897929 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7087e292-a9fd-4420-bb4a-eb2d50355082-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.952781 4573 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.952807 4573 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.952818 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kzkl\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-kube-api-access-5kzkl\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.952856 4573 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.952868 4573 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7087e292-a9fd-4420-bb4a-eb2d50355082-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.952877 4573 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.952891 4573 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7087e292-a9fd-4420-bb4a-eb2d50355082-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.957796 4573 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.958038 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-server-conf" (OuterVolumeSpecName: "server-conf") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.967613 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-config-data" (OuterVolumeSpecName: "config-data") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:32 crc kubenswrapper[4573]: I1203 09:05:32.992238 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.028348 4573 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.060035 4573 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.066224 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.066243 4573 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7087e292-a9fd-4420-bb4a-eb2d50355082-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.167944 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9988868b-f38f-4e95-af61-1834e0e67a8b-erlang-cookie-secret\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168078 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-plugins-conf\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168115 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-plugins\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168159 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6tm9\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-kube-api-access-h6tm9\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168182 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-server-conf\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168217 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-tls\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168269 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168313 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-confd\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168332 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-config-data\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168372 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9988868b-f38f-4e95-af61-1834e0e67a8b-pod-info\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.168549 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-erlang-cookie\") pod \"9988868b-f38f-4e95-af61-1834e0e67a8b\" (UID: \"9988868b-f38f-4e95-af61-1834e0e67a8b\") " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.169088 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7087e292-a9fd-4420-bb4a-eb2d50355082" (UID: "7087e292-a9fd-4420-bb4a-eb2d50355082"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.174572 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.179950 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.180457 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.181240 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.191031 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.197115 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9988868b-f38f-4e95-af61-1834e0e67a8b-pod-info" (OuterVolumeSpecName: "pod-info") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.198272 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-kube-api-access-h6tm9" (OuterVolumeSpecName: "kube-api-access-h6tm9") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "kube-api-access-h6tm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.210235 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9988868b-f38f-4e95-af61-1834e0e67a8b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.273329 4573 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7087e292-a9fd-4420-bb4a-eb2d50355082-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.273477 4573 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.273604 4573 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9988868b-f38f-4e95-af61-1834e0e67a8b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.273681 4573 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.273753 4573 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.273822 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6tm9\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-kube-api-access-h6tm9\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.273874 4573 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.273937 4573 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.274000 4573 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9988868b-f38f-4e95-af61-1834e0e67a8b-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.298062 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-config-data" (OuterVolumeSpecName: "config-data") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.376094 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.392255 4573 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.416495 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-server-conf" (OuterVolumeSpecName: "server-conf") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.429691 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9988868b-f38f-4e95-af61-1834e0e67a8b" (UID: "9988868b-f38f-4e95-af61-1834e0e67a8b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.477593 4573 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9988868b-f38f-4e95-af61-1834e0e67a8b-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.477634 4573 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.477643 4573 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9988868b-f38f-4e95-af61-1834e0e67a8b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.600652 4573 generic.go:334] "Generic (PLEG): container finished" podID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerID="5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64" exitCode=0 Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.600891 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9988868b-f38f-4e95-af61-1834e0e67a8b","Type":"ContainerDied","Data":"5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64"} Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.600971 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9988868b-f38f-4e95-af61-1834e0e67a8b","Type":"ContainerDied","Data":"a998b66be8a88486c735874ab97dfbe2f20f83669597b82a6223fde37b4da589"} Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.600986 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.600995 4573 scope.go:117] "RemoveContainer" containerID="5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.601083 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.642039 4573 scope.go:117] "RemoveContainer" containerID="b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.657307 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.707120 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.718960 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.730669 4573 scope.go:117] "RemoveContainer" containerID="5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64" Dec 03 09:05:33 crc kubenswrapper[4573]: E1203 09:05:33.732648 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64\": container with ID starting with 5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64 not found: ID does not exist" containerID="5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.732739 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64"} err="failed to get container status \"5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64\": rpc error: code = NotFound desc = could not find container \"5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64\": container with ID starting with 5db53c8d1fb4eaf4289440e3a074d7ad54001e92fb518e16d852b87c3e770e64 not found: ID does not exist" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.732812 4573 scope.go:117] "RemoveContainer" containerID="b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.732912 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 09:05:33 crc kubenswrapper[4573]: E1203 09:05:33.736328 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca\": container with ID starting with b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca not found: ID does not exist" containerID="b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.736412 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca"} err="failed to get container status \"b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca\": rpc error: code = NotFound desc = could not find container \"b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca\": container with ID starting with b9f2131bf9cc626bb4d78bf970c6fcc8d530d3ef2dd12c72aa3915dffcbb7aca not found: ID does not exist" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.758663 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 09:05:33 crc kubenswrapper[4573]: E1203 09:05:33.759234 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerName="rabbitmq" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.759254 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerName="rabbitmq" Dec 03 09:05:33 crc kubenswrapper[4573]: E1203 09:05:33.759270 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerName="setup-container" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.759277 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerName="setup-container" Dec 03 09:05:33 crc kubenswrapper[4573]: E1203 09:05:33.759291 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerName="setup-container" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.759297 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerName="setup-container" Dec 03 09:05:33 crc kubenswrapper[4573]: E1203 09:05:33.759307 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerName="rabbitmq" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.759313 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerName="rabbitmq" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.759538 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7087e292-a9fd-4420-bb4a-eb2d50355082" containerName="rabbitmq" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.759563 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="9988868b-f38f-4e95-af61-1834e0e67a8b" containerName="rabbitmq" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.760672 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.765382 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.765402 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.765614 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.765671 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-fzd4g" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.765749 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.765829 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.765885 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.786541 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.814755 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.816516 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.828426 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.828653 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.828794 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.828958 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hb2p4" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.829126 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.829225 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.829249 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.834086 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.895330 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2bqp\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-kube-api-access-g2bqp\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.895385 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.895418 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.895727 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.895792 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.895859 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70fb1808-0ac8-4096-abff-76c44b1245f8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.895882 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.896099 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-config-data\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.896224 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.896347 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.896406 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70fb1808-0ac8-4096-abff-76c44b1245f8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998029 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-config-data\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998124 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f038afb-2485-4e6e-8e0c-99ec178f4268-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998149 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998170 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998211 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998234 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfmbd\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-kube-api-access-kfmbd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998255 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70fb1808-0ac8-4096-abff-76c44b1245f8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998472 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998517 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2bqp\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-kube-api-access-g2bqp\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998543 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998568 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998604 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f038afb-2485-4e6e-8e0c-99ec178f4268-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998626 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998652 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998678 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998708 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998737 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998767 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998825 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998846 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998889 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998912 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70fb1808-0ac8-4096-abff-76c44b1245f8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.998972 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.999191 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-config-data\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.999331 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.999654 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 03 09:05:33 crc kubenswrapper[4573]: I1203 09:05:33.999792 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.005385 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/70fb1808-0ac8-4096-abff-76c44b1245f8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.012263 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.012713 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/70fb1808-0ac8-4096-abff-76c44b1245f8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.012253 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.013305 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/70fb1808-0ac8-4096-abff-76c44b1245f8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.022582 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2bqp\" (UniqueName: \"kubernetes.io/projected/70fb1808-0ac8-4096-abff-76c44b1245f8-kube-api-access-g2bqp\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.042263 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7087e292-a9fd-4420-bb4a-eb2d50355082" path="/var/lib/kubelet/pods/7087e292-a9fd-4420-bb4a-eb2d50355082/volumes" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.044018 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9988868b-f38f-4e95-af61-1834e0e67a8b" path="/var/lib/kubelet/pods/9988868b-f38f-4e95-af61-1834e0e67a8b/volumes" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.044597 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"70fb1808-0ac8-4096-abff-76c44b1245f8\") " pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.086866 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.101654 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f038afb-2485-4e6e-8e0c-99ec178f4268-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.101722 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.101837 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfmbd\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-kube-api-access-kfmbd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.101943 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.102001 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f038afb-2485-4e6e-8e0c-99ec178f4268-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.102023 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.102073 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.102116 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.102162 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.102197 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.102247 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.106631 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.106685 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f038afb-2485-4e6e-8e0c-99ec178f4268-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.107310 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.107647 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.107884 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.108356 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.108795 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f038afb-2485-4e6e-8e0c-99ec178f4268-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.108872 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.110602 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f038afb-2485-4e6e-8e0c-99ec178f4268-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.113080 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.130443 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfmbd\" (UniqueName: \"kubernetes.io/projected/5f038afb-2485-4e6e-8e0c-99ec178f4268-kube-api-access-kfmbd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.142293 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5f038afb-2485-4e6e-8e0c-99ec178f4268\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.158377 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.607218 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 09:05:34 crc kubenswrapper[4573]: I1203 09:05:34.699942 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 09:05:35 crc kubenswrapper[4573]: I1203 09:05:35.648218 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5f038afb-2485-4e6e-8e0c-99ec178f4268","Type":"ContainerStarted","Data":"0c1bac611158c4b79e71c9c0505324b2b0666a6cbadb8db8073ae267dc8dfa33"} Dec 03 09:05:35 crc kubenswrapper[4573]: I1203 09:05:35.649797 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70fb1808-0ac8-4096-abff-76c44b1245f8","Type":"ContainerStarted","Data":"06fae5ed2eef0436bfff39c1273b67344f16c5f4e5c50dd1fb7c8d9ea2ed0b14"} Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.249734 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mw728"] Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.251581 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.254175 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.270887 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mw728"] Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.431241 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-config\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.431348 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.431469 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.431524 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.431555 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djt56\" (UniqueName: \"kubernetes.io/projected/3169986e-2cf3-4492-8fac-c41b087aa59b-kube-api-access-djt56\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.431580 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.431601 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-svc\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.533186 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.533652 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.533778 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.533862 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djt56\" (UniqueName: \"kubernetes.io/projected/3169986e-2cf3-4492-8fac-c41b087aa59b-kube-api-access-djt56\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.533939 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.534030 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-svc\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.534361 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.534653 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.534660 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.534850 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.535187 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-svc\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.535381 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-config\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.536016 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-config\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.553374 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djt56\" (UniqueName: \"kubernetes.io/projected/3169986e-2cf3-4492-8fac-c41b087aa59b-kube-api-access-djt56\") pod \"dnsmasq-dns-d558885bc-mw728\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.578651 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.672660 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5f038afb-2485-4e6e-8e0c-99ec178f4268","Type":"ContainerStarted","Data":"17c6236449177c912782e782dec54b4d947434c621580e93c22284723ad4eb7d"} Dec 03 09:05:36 crc kubenswrapper[4573]: I1203 09:05:36.682296 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70fb1808-0ac8-4096-abff-76c44b1245f8","Type":"ContainerStarted","Data":"4c202dd3fda662c000b6bf34d33e5ef498bd821e8d15f3c3c09d165cbc34e2c2"} Dec 03 09:05:37 crc kubenswrapper[4573]: I1203 09:05:37.302723 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mw728"] Dec 03 09:05:37 crc kubenswrapper[4573]: I1203 09:05:37.709205 4573 generic.go:334] "Generic (PLEG): container finished" podID="3169986e-2cf3-4492-8fac-c41b087aa59b" containerID="1640cc9ca27ee16ec84d4946005346391e9ae2466beebff03099201cdfc01a8a" exitCode=0 Dec 03 09:05:37 crc kubenswrapper[4573]: I1203 09:05:37.709405 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mw728" event={"ID":"3169986e-2cf3-4492-8fac-c41b087aa59b","Type":"ContainerDied","Data":"1640cc9ca27ee16ec84d4946005346391e9ae2466beebff03099201cdfc01a8a"} Dec 03 09:05:37 crc kubenswrapper[4573]: I1203 09:05:37.709668 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mw728" event={"ID":"3169986e-2cf3-4492-8fac-c41b087aa59b","Type":"ContainerStarted","Data":"eb93cb4dd5c566cf8ba14c63e73001d6d6a4707b40a7e4af9d36a85be9318c57"} Dec 03 09:05:38 crc kubenswrapper[4573]: I1203 09:05:38.720188 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mw728" event={"ID":"3169986e-2cf3-4492-8fac-c41b087aa59b","Type":"ContainerStarted","Data":"9df0c3b7bc7aeb79e2e39395ac3d8c5cf1553b7ba5a2ca010b20e65ea9dd2433"} Dec 03 09:05:38 crc kubenswrapper[4573]: I1203 09:05:38.720482 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:46 crc kubenswrapper[4573]: I1203 09:05:46.581368 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:46 crc kubenswrapper[4573]: I1203 09:05:46.611759 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-mw728" podStartSLOduration=10.611731362 podStartE2EDuration="10.611731362s" podCreationTimestamp="2025-12-03 09:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:05:38.746959845 +0000 UTC m=+1659.315339124" watchObservedRunningTime="2025-12-03 09:05:46.611731362 +0000 UTC m=+1667.180110621" Dec 03 09:05:46 crc kubenswrapper[4573]: I1203 09:05:46.721022 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-h955r"] Dec 03 09:05:46 crc kubenswrapper[4573]: I1203 09:05:46.721499 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" podUID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" containerName="dnsmasq-dns" containerID="cri-o://64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e" gracePeriod=10 Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.115336 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d47554775-5hw5x"] Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.119035 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.138374 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d47554775-5hw5x"] Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.191598 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-config\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.191657 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-dns-swift-storage-0\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.191691 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvfqf\" (UniqueName: \"kubernetes.io/projected/eb528c46-bf42-4870-850c-dcf96e5a59b9-kube-api-access-gvfqf\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.191719 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.191768 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.191797 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-dns-svc\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.191874 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-ovsdbserver-nb\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.293968 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.294067 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-dns-svc\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.294175 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-ovsdbserver-nb\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.294213 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-config\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.294238 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-dns-swift-storage-0\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.294271 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvfqf\" (UniqueName: \"kubernetes.io/projected/eb528c46-bf42-4870-850c-dcf96e5a59b9-kube-api-access-gvfqf\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.294303 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.295204 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-ovsdbserver-sb\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.295244 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-openstack-edpm-ipam\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.295918 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-dns-svc\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.296025 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-dns-swift-storage-0\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.296140 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-config\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.296578 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb528c46-bf42-4870-850c-dcf96e5a59b9-ovsdbserver-nb\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.321328 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvfqf\" (UniqueName: \"kubernetes.io/projected/eb528c46-bf42-4870-850c-dcf96e5a59b9-kube-api-access-gvfqf\") pod \"dnsmasq-dns-6d47554775-5hw5x\" (UID: \"eb528c46-bf42-4870-850c-dcf96e5a59b9\") " pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.381190 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.396145 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-svc\") pod \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.396337 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-sb\") pod \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.396365 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2ptv\" (UniqueName: \"kubernetes.io/projected/4082d0f4-8d86-4382-bcb6-fe560d365ae3-kube-api-access-p2ptv\") pod \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.396389 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-config\") pod \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.403583 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4082d0f4-8d86-4382-bcb6-fe560d365ae3-kube-api-access-p2ptv" (OuterVolumeSpecName: "kube-api-access-p2ptv") pod "4082d0f4-8d86-4382-bcb6-fe560d365ae3" (UID: "4082d0f4-8d86-4382-bcb6-fe560d365ae3"). InnerVolumeSpecName "kube-api-access-p2ptv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.453885 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.497932 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-config" (OuterVolumeSpecName: "config") pod "4082d0f4-8d86-4382-bcb6-fe560d365ae3" (UID: "4082d0f4-8d86-4382-bcb6-fe560d365ae3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.498590 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-config\") pod \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.498759 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-swift-storage-0\") pod \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.498869 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-nb\") pod \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\" (UID: \"4082d0f4-8d86-4382-bcb6-fe560d365ae3\") " Dec 03 09:05:47 crc kubenswrapper[4573]: W1203 09:05:47.499554 4573 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/4082d0f4-8d86-4382-bcb6-fe560d365ae3/volumes/kubernetes.io~configmap/config Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.499943 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-config" (OuterVolumeSpecName: "config") pod "4082d0f4-8d86-4382-bcb6-fe560d365ae3" (UID: "4082d0f4-8d86-4382-bcb6-fe560d365ae3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.501422 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2ptv\" (UniqueName: \"kubernetes.io/projected/4082d0f4-8d86-4382-bcb6-fe560d365ae3-kube-api-access-p2ptv\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.501445 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.525330 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4082d0f4-8d86-4382-bcb6-fe560d365ae3" (UID: "4082d0f4-8d86-4382-bcb6-fe560d365ae3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.538697 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4082d0f4-8d86-4382-bcb6-fe560d365ae3" (UID: "4082d0f4-8d86-4382-bcb6-fe560d365ae3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.585463 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4082d0f4-8d86-4382-bcb6-fe560d365ae3" (UID: "4082d0f4-8d86-4382-bcb6-fe560d365ae3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.605985 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4082d0f4-8d86-4382-bcb6-fe560d365ae3" (UID: "4082d0f4-8d86-4382-bcb6-fe560d365ae3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.619723 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.619808 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.619821 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.619831 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4082d0f4-8d86-4382-bcb6-fe560d365ae3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:47 crc kubenswrapper[4573]: W1203 09:05:47.791720 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb528c46_bf42_4870_850c_dcf96e5a59b9.slice/crio-6cd02970725b22fe6dfd25971ddeba7b3d9b5b3ba5c4d50fdd4304b1ada08f91 WatchSource:0}: Error finding container 6cd02970725b22fe6dfd25971ddeba7b3d9b5b3ba5c4d50fdd4304b1ada08f91: Status 404 returned error can't find the container with id 6cd02970725b22fe6dfd25971ddeba7b3d9b5b3ba5c4d50fdd4304b1ada08f91 Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.799177 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d47554775-5hw5x"] Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.870000 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d47554775-5hw5x" event={"ID":"eb528c46-bf42-4870-850c-dcf96e5a59b9","Type":"ContainerStarted","Data":"6cd02970725b22fe6dfd25971ddeba7b3d9b5b3ba5c4d50fdd4304b1ada08f91"} Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.872638 4573 generic.go:334] "Generic (PLEG): container finished" podID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" containerID="64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e" exitCode=0 Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.872664 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" event={"ID":"4082d0f4-8d86-4382-bcb6-fe560d365ae3","Type":"ContainerDied","Data":"64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e"} Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.872683 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" event={"ID":"4082d0f4-8d86-4382-bcb6-fe560d365ae3","Type":"ContainerDied","Data":"9ade5d4d4435801d78cf2a099c8f6ce13117fec395ce8458908e3b2c98c464b0"} Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.872702 4573 scope.go:117] "RemoveContainer" containerID="64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.872717 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-h955r" Dec 03 09:05:47 crc kubenswrapper[4573]: I1203 09:05:47.989869 4573 scope.go:117] "RemoveContainer" containerID="a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35" Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.015722 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-h955r"] Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.027993 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-h955r"] Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.049523 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" path="/var/lib/kubelet/pods/4082d0f4-8d86-4382-bcb6-fe560d365ae3/volumes" Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.050956 4573 scope.go:117] "RemoveContainer" containerID="64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e" Dec 03 09:05:48 crc kubenswrapper[4573]: E1203 09:05:48.053070 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e\": container with ID starting with 64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e not found: ID does not exist" containerID="64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e" Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.053113 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e"} err="failed to get container status \"64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e\": rpc error: code = NotFound desc = could not find container \"64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e\": container with ID starting with 64285ef43fab2d4a329d12d61999de0b19aab1ca5e3bc09bedb6acde0980951e not found: ID does not exist" Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.053142 4573 scope.go:117] "RemoveContainer" containerID="a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35" Dec 03 09:05:48 crc kubenswrapper[4573]: E1203 09:05:48.053797 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35\": container with ID starting with a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35 not found: ID does not exist" containerID="a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35" Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.053831 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35"} err="failed to get container status \"a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35\": rpc error: code = NotFound desc = could not find container \"a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35\": container with ID starting with a7973e470e11faa4cf5f13bb7e95831be8254bbda182668d07675f65de4b6c35 not found: ID does not exist" Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.886820 4573 generic.go:334] "Generic (PLEG): container finished" podID="eb528c46-bf42-4870-850c-dcf96e5a59b9" containerID="8373d052398c41941cbf81594a81b39a0cdc0eca8247a1b62ee939752ae9765a" exitCode=0 Dec 03 09:05:48 crc kubenswrapper[4573]: I1203 09:05:48.886934 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d47554775-5hw5x" event={"ID":"eb528c46-bf42-4870-850c-dcf96e5a59b9","Type":"ContainerDied","Data":"8373d052398c41941cbf81594a81b39a0cdc0eca8247a1b62ee939752ae9765a"} Dec 03 09:05:49 crc kubenswrapper[4573]: I1203 09:05:49.902869 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d47554775-5hw5x" event={"ID":"eb528c46-bf42-4870-850c-dcf96e5a59b9","Type":"ContainerStarted","Data":"4ab47eb724fe7a02d7816e7653e9b698b991576c3f0f4f4fb816d06c7ab49238"} Dec 03 09:05:49 crc kubenswrapper[4573]: I1203 09:05:49.903348 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:49 crc kubenswrapper[4573]: I1203 09:05:49.934929 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d47554775-5hw5x" podStartSLOduration=2.934908162 podStartE2EDuration="2.934908162s" podCreationTimestamp="2025-12-03 09:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:05:49.92646605 +0000 UTC m=+1670.494845309" watchObservedRunningTime="2025-12-03 09:05:49.934908162 +0000 UTC m=+1670.503287421" Dec 03 09:05:56 crc kubenswrapper[4573]: I1203 09:05:56.942893 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:05:56 crc kubenswrapper[4573]: I1203 09:05:56.943279 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:05:56 crc kubenswrapper[4573]: I1203 09:05:56.943342 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:05:56 crc kubenswrapper[4573]: I1203 09:05:56.944801 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:05:56 crc kubenswrapper[4573]: I1203 09:05:56.944885 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" gracePeriod=600 Dec 03 09:05:57 crc kubenswrapper[4573]: E1203 09:05:57.076471 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:05:57 crc kubenswrapper[4573]: I1203 09:05:57.455810 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d47554775-5hw5x" Dec 03 09:05:57 crc kubenswrapper[4573]: I1203 09:05:57.548318 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mw728"] Dec 03 09:05:57 crc kubenswrapper[4573]: I1203 09:05:57.548656 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-mw728" podUID="3169986e-2cf3-4492-8fac-c41b087aa59b" containerName="dnsmasq-dns" containerID="cri-o://9df0c3b7bc7aeb79e2e39395ac3d8c5cf1553b7ba5a2ca010b20e65ea9dd2433" gracePeriod=10 Dec 03 09:05:57 crc kubenswrapper[4573]: I1203 09:05:57.991461 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" exitCode=0 Dec 03 09:05:57 crc kubenswrapper[4573]: I1203 09:05:57.991676 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85"} Dec 03 09:05:57 crc kubenswrapper[4573]: I1203 09:05:57.991802 4573 scope.go:117] "RemoveContainer" containerID="2b3b734897ed0efe44e3ec39a1d49da70613f4f9958b9ef5aee9f0572f629605" Dec 03 09:05:57 crc kubenswrapper[4573]: I1203 09:05:57.992822 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:05:57 crc kubenswrapper[4573]: E1203 09:05:57.993580 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.000032 4573 generic.go:334] "Generic (PLEG): container finished" podID="3169986e-2cf3-4492-8fac-c41b087aa59b" containerID="9df0c3b7bc7aeb79e2e39395ac3d8c5cf1553b7ba5a2ca010b20e65ea9dd2433" exitCode=0 Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.000139 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mw728" event={"ID":"3169986e-2cf3-4492-8fac-c41b087aa59b","Type":"ContainerDied","Data":"9df0c3b7bc7aeb79e2e39395ac3d8c5cf1553b7ba5a2ca010b20e65ea9dd2433"} Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.114630 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.309704 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-svc\") pod \"3169986e-2cf3-4492-8fac-c41b087aa59b\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.309813 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-sb\") pod \"3169986e-2cf3-4492-8fac-c41b087aa59b\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.309842 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-swift-storage-0\") pod \"3169986e-2cf3-4492-8fac-c41b087aa59b\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.309908 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-nb\") pod \"3169986e-2cf3-4492-8fac-c41b087aa59b\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.309932 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-openstack-edpm-ipam\") pod \"3169986e-2cf3-4492-8fac-c41b087aa59b\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.310077 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-config\") pod \"3169986e-2cf3-4492-8fac-c41b087aa59b\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.310166 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djt56\" (UniqueName: \"kubernetes.io/projected/3169986e-2cf3-4492-8fac-c41b087aa59b-kube-api-access-djt56\") pod \"3169986e-2cf3-4492-8fac-c41b087aa59b\" (UID: \"3169986e-2cf3-4492-8fac-c41b087aa59b\") " Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.321296 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3169986e-2cf3-4492-8fac-c41b087aa59b-kube-api-access-djt56" (OuterVolumeSpecName: "kube-api-access-djt56") pod "3169986e-2cf3-4492-8fac-c41b087aa59b" (UID: "3169986e-2cf3-4492-8fac-c41b087aa59b"). InnerVolumeSpecName "kube-api-access-djt56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.407758 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-config" (OuterVolumeSpecName: "config") pod "3169986e-2cf3-4492-8fac-c41b087aa59b" (UID: "3169986e-2cf3-4492-8fac-c41b087aa59b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.415356 4573 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.415412 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djt56\" (UniqueName: \"kubernetes.io/projected/3169986e-2cf3-4492-8fac-c41b087aa59b-kube-api-access-djt56\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.426742 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3169986e-2cf3-4492-8fac-c41b087aa59b" (UID: "3169986e-2cf3-4492-8fac-c41b087aa59b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.430708 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "3169986e-2cf3-4492-8fac-c41b087aa59b" (UID: "3169986e-2cf3-4492-8fac-c41b087aa59b"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.436650 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3169986e-2cf3-4492-8fac-c41b087aa59b" (UID: "3169986e-2cf3-4492-8fac-c41b087aa59b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.474738 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3169986e-2cf3-4492-8fac-c41b087aa59b" (UID: "3169986e-2cf3-4492-8fac-c41b087aa59b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.477893 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3169986e-2cf3-4492-8fac-c41b087aa59b" (UID: "3169986e-2cf3-4492-8fac-c41b087aa59b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.517663 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.517710 4573 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.517725 4573 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.517734 4573 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:58 crc kubenswrapper[4573]: I1203 09:05:58.517743 4573 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3169986e-2cf3-4492-8fac-c41b087aa59b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:05:59 crc kubenswrapper[4573]: I1203 09:05:59.020908 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-mw728" event={"ID":"3169986e-2cf3-4492-8fac-c41b087aa59b","Type":"ContainerDied","Data":"eb93cb4dd5c566cf8ba14c63e73001d6d6a4707b40a7e4af9d36a85be9318c57"} Dec 03 09:05:59 crc kubenswrapper[4573]: I1203 09:05:59.021198 4573 scope.go:117] "RemoveContainer" containerID="9df0c3b7bc7aeb79e2e39395ac3d8c5cf1553b7ba5a2ca010b20e65ea9dd2433" Dec 03 09:05:59 crc kubenswrapper[4573]: I1203 09:05:59.024132 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-mw728" Dec 03 09:05:59 crc kubenswrapper[4573]: I1203 09:05:59.057209 4573 scope.go:117] "RemoveContainer" containerID="1640cc9ca27ee16ec84d4946005346391e9ae2466beebff03099201cdfc01a8a" Dec 03 09:05:59 crc kubenswrapper[4573]: I1203 09:05:59.074226 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mw728"] Dec 03 09:05:59 crc kubenswrapper[4573]: I1203 09:05:59.085606 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-mw728"] Dec 03 09:06:00 crc kubenswrapper[4573]: I1203 09:06:00.049679 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3169986e-2cf3-4492-8fac-c41b087aa59b" path="/var/lib/kubelet/pods/3169986e-2cf3-4492-8fac-c41b087aa59b/volumes" Dec 03 09:06:04 crc kubenswrapper[4573]: I1203 09:06:04.689535 4573 scope.go:117] "RemoveContainer" containerID="823decdf41f2767a07f9b8a0f91c06403b0695e8cd0d75c5f5350ba36ed7b5c9" Dec 03 09:06:04 crc kubenswrapper[4573]: I1203 09:06:04.719481 4573 scope.go:117] "RemoveContainer" containerID="8fc49a0e39d5014a0145f26bd4d0be966ee9fad4deaaf9c035afcb09e6c961b0" Dec 03 09:06:04 crc kubenswrapper[4573]: I1203 09:06:04.771378 4573 scope.go:117] "RemoveContainer" containerID="c5991229e5c00013cea09dc1f612403384de8e841bcf82e821749bd6642ea1a4" Dec 03 09:06:04 crc kubenswrapper[4573]: I1203 09:06:04.816423 4573 scope.go:117] "RemoveContainer" containerID="5f27bfda26561875fed220674d65c66eac1bdc140eca3aa513d193416a640c55" Dec 03 09:06:04 crc kubenswrapper[4573]: I1203 09:06:04.850930 4573 scope.go:117] "RemoveContainer" containerID="f38acc7f0c62619582426dcd3c2f17ed18d6c6e8c070e0ddb5862554c5274a5e" Dec 03 09:06:04 crc kubenswrapper[4573]: I1203 09:06:04.904040 4573 scope.go:117] "RemoveContainer" containerID="dd14c68ccd8e3b92c3b298436c0b7a83d1beb287c50d190663b4e748a4d61964" Dec 03 09:06:04 crc kubenswrapper[4573]: I1203 09:06:04.949777 4573 scope.go:117] "RemoveContainer" containerID="6ef8e5934c94595f86df204e3ac8183e89565374ccead7c035e8695fe6a1b12c" Dec 03 09:06:09 crc kubenswrapper[4573]: I1203 09:06:09.162021 4573 generic.go:334] "Generic (PLEG): container finished" podID="5f038afb-2485-4e6e-8e0c-99ec178f4268" containerID="17c6236449177c912782e782dec54b4d947434c621580e93c22284723ad4eb7d" exitCode=0 Dec 03 09:06:09 crc kubenswrapper[4573]: I1203 09:06:09.162122 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5f038afb-2485-4e6e-8e0c-99ec178f4268","Type":"ContainerDied","Data":"17c6236449177c912782e782dec54b4d947434c621580e93c22284723ad4eb7d"} Dec 03 09:06:09 crc kubenswrapper[4573]: I1203 09:06:09.172777 4573 generic.go:334] "Generic (PLEG): container finished" podID="70fb1808-0ac8-4096-abff-76c44b1245f8" containerID="4c202dd3fda662c000b6bf34d33e5ef498bd821e8d15f3c3c09d165cbc34e2c2" exitCode=0 Dec 03 09:06:09 crc kubenswrapper[4573]: I1203 09:06:09.172845 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70fb1808-0ac8-4096-abff-76c44b1245f8","Type":"ContainerDied","Data":"4c202dd3fda662c000b6bf34d33e5ef498bd821e8d15f3c3c09d165cbc34e2c2"} Dec 03 09:06:10 crc kubenswrapper[4573]: I1203 09:06:10.061215 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:06:10 crc kubenswrapper[4573]: E1203 09:06:10.062263 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:06:10 crc kubenswrapper[4573]: I1203 09:06:10.216643 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"70fb1808-0ac8-4096-abff-76c44b1245f8","Type":"ContainerStarted","Data":"c022c27b709682f0397f818128036e995d06916b5d2950a01095af8ed425bf3c"} Dec 03 09:06:10 crc kubenswrapper[4573]: I1203 09:06:10.218592 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 09:06:10 crc kubenswrapper[4573]: I1203 09:06:10.235788 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5f038afb-2485-4e6e-8e0c-99ec178f4268","Type":"ContainerStarted","Data":"75026451373cffd04a3df4dc9ac8fcb415dcc49a8b1812ce7fe886783c022eb1"} Dec 03 09:06:10 crc kubenswrapper[4573]: I1203 09:06:10.236132 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:06:10 crc kubenswrapper[4573]: I1203 09:06:10.262673 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.262632404 podStartE2EDuration="37.262632404s" podCreationTimestamp="2025-12-03 09:05:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:06:10.24610962 +0000 UTC m=+1690.814488879" watchObservedRunningTime="2025-12-03 09:06:10.262632404 +0000 UTC m=+1690.831011663" Dec 03 09:06:10 crc kubenswrapper[4573]: I1203 09:06:10.312655 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.312623097 podStartE2EDuration="37.312623097s" podCreationTimestamp="2025-12-03 09:05:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:06:10.289664436 +0000 UTC m=+1690.858043695" watchObservedRunningTime="2025-12-03 09:06:10.312623097 +0000 UTC m=+1690.881002356" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.601975 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7"] Dec 03 09:06:16 crc kubenswrapper[4573]: E1203 09:06:16.603969 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3169986e-2cf3-4492-8fac-c41b087aa59b" containerName="init" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.603993 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3169986e-2cf3-4492-8fac-c41b087aa59b" containerName="init" Dec 03 09:06:16 crc kubenswrapper[4573]: E1203 09:06:16.604039 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" containerName="init" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.604073 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" containerName="init" Dec 03 09:06:16 crc kubenswrapper[4573]: E1203 09:06:16.604105 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" containerName="dnsmasq-dns" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.604112 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" containerName="dnsmasq-dns" Dec 03 09:06:16 crc kubenswrapper[4573]: E1203 09:06:16.604134 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3169986e-2cf3-4492-8fac-c41b087aa59b" containerName="dnsmasq-dns" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.604143 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3169986e-2cf3-4492-8fac-c41b087aa59b" containerName="dnsmasq-dns" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.604413 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3169986e-2cf3-4492-8fac-c41b087aa59b" containerName="dnsmasq-dns" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.604433 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="4082d0f4-8d86-4382-bcb6-fe560d365ae3" containerName="dnsmasq-dns" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.605587 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.609121 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.609351 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.610093 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.610323 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.639386 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7"] Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.706740 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.706841 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skqv7\" (UniqueName: \"kubernetes.io/projected/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-kube-api-access-skqv7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.707153 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.707239 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.809086 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.809208 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.809266 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skqv7\" (UniqueName: \"kubernetes.io/projected/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-kube-api-access-skqv7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.809330 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.818944 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.821510 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.832855 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skqv7\" (UniqueName: \"kubernetes.io/projected/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-kube-api-access-skqv7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.842001 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:16 crc kubenswrapper[4573]: I1203 09:06:16.931867 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:17 crc kubenswrapper[4573]: I1203 09:06:17.618266 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7"] Dec 03 09:06:18 crc kubenswrapper[4573]: I1203 09:06:18.348853 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" event={"ID":"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74","Type":"ContainerStarted","Data":"a58aa129945dc0377d1f2d912c0908ca041fa4a2fef942a5259a55a4635467a9"} Dec 03 09:06:24 crc kubenswrapper[4573]: I1203 09:06:24.089490 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 09:06:24 crc kubenswrapper[4573]: I1203 09:06:24.164274 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 09:06:25 crc kubenswrapper[4573]: I1203 09:06:25.030815 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:06:25 crc kubenswrapper[4573]: E1203 09:06:25.031460 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:06:33 crc kubenswrapper[4573]: I1203 09:06:33.737896 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" event={"ID":"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74","Type":"ContainerStarted","Data":"9b8f9f0fd85992d3cb746d2245418148ddedd0d58654ed5da166aabe3ecfd02b"} Dec 03 09:06:33 crc kubenswrapper[4573]: I1203 09:06:33.765871 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" podStartSLOduration=2.522926292 podStartE2EDuration="17.765844485s" podCreationTimestamp="2025-12-03 09:06:16 +0000 UTC" firstStartedPulling="2025-12-03 09:06:17.631365627 +0000 UTC m=+1698.199744886" lastFinishedPulling="2025-12-03 09:06:32.87428381 +0000 UTC m=+1713.442663079" observedRunningTime="2025-12-03 09:06:33.753873286 +0000 UTC m=+1714.322252555" watchObservedRunningTime="2025-12-03 09:06:33.765844485 +0000 UTC m=+1714.334223754" Dec 03 09:06:38 crc kubenswrapper[4573]: I1203 09:06:38.030878 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:06:38 crc kubenswrapper[4573]: E1203 09:06:38.032780 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:06:44 crc kubenswrapper[4573]: I1203 09:06:44.869467 4573 generic.go:334] "Generic (PLEG): container finished" podID="dd487e4c-80de-48b3-aaeb-1a3c9d5eda74" containerID="9b8f9f0fd85992d3cb746d2245418148ddedd0d58654ed5da166aabe3ecfd02b" exitCode=0 Dec 03 09:06:44 crc kubenswrapper[4573]: I1203 09:06:44.869596 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" event={"ID":"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74","Type":"ContainerDied","Data":"9b8f9f0fd85992d3cb746d2245418148ddedd0d58654ed5da166aabe3ecfd02b"} Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.409785 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.454783 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skqv7\" (UniqueName: \"kubernetes.io/projected/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-kube-api-access-skqv7\") pod \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.454997 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-repo-setup-combined-ca-bundle\") pod \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.455062 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-inventory\") pod \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.455516 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-ssh-key\") pod \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\" (UID: \"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74\") " Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.464788 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "dd487e4c-80de-48b3-aaeb-1a3c9d5eda74" (UID: "dd487e4c-80de-48b3-aaeb-1a3c9d5eda74"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.465017 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-kube-api-access-skqv7" (OuterVolumeSpecName: "kube-api-access-skqv7") pod "dd487e4c-80de-48b3-aaeb-1a3c9d5eda74" (UID: "dd487e4c-80de-48b3-aaeb-1a3c9d5eda74"). InnerVolumeSpecName "kube-api-access-skqv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.490208 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-inventory" (OuterVolumeSpecName: "inventory") pod "dd487e4c-80de-48b3-aaeb-1a3c9d5eda74" (UID: "dd487e4c-80de-48b3-aaeb-1a3c9d5eda74"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.514240 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dd487e4c-80de-48b3-aaeb-1a3c9d5eda74" (UID: "dd487e4c-80de-48b3-aaeb-1a3c9d5eda74"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.558947 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.558986 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skqv7\" (UniqueName: \"kubernetes.io/projected/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-kube-api-access-skqv7\") on node \"crc\" DevicePath \"\"" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.558998 4573 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.559008 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd487e4c-80de-48b3-aaeb-1a3c9d5eda74-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.901749 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" event={"ID":"dd487e4c-80de-48b3-aaeb-1a3c9d5eda74","Type":"ContainerDied","Data":"a58aa129945dc0377d1f2d912c0908ca041fa4a2fef942a5259a55a4635467a9"} Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.902311 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a58aa129945dc0377d1f2d912c0908ca041fa4a2fef942a5259a55a4635467a9" Dec 03 09:06:46 crc kubenswrapper[4573]: I1203 09:06:46.901842 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.035136 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc"] Dec 03 09:06:47 crc kubenswrapper[4573]: E1203 09:06:47.036129 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd487e4c-80de-48b3-aaeb-1a3c9d5eda74" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.036178 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd487e4c-80de-48b3-aaeb-1a3c9d5eda74" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.036630 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd487e4c-80de-48b3-aaeb-1a3c9d5eda74" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.038777 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.042369 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.048888 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.050851 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc"] Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.051726 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.052074 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.069632 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.069676 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz6rk\" (UniqueName: \"kubernetes.io/projected/609d7944-6cf6-47c4-a1f2-2437c57703d2-kube-api-access-gz6rk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.069757 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.171791 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.171966 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.171994 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz6rk\" (UniqueName: \"kubernetes.io/projected/609d7944-6cf6-47c4-a1f2-2437c57703d2-kube-api-access-gz6rk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.177864 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.178389 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.199004 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz6rk\" (UniqueName: \"kubernetes.io/projected/609d7944-6cf6-47c4-a1f2-2437c57703d2-kube-api-access-gz6rk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-lbctc\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:47 crc kubenswrapper[4573]: I1203 09:06:47.385087 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:48 crc kubenswrapper[4573]: I1203 09:06:48.015934 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc"] Dec 03 09:06:48 crc kubenswrapper[4573]: W1203 09:06:48.023210 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod609d7944_6cf6_47c4_a1f2_2437c57703d2.slice/crio-4f484ac1378459a570a1e2a50bb27edba8176e69f87d0ec5cf386ed6867d4bf6 WatchSource:0}: Error finding container 4f484ac1378459a570a1e2a50bb27edba8176e69f87d0ec5cf386ed6867d4bf6: Status 404 returned error can't find the container with id 4f484ac1378459a570a1e2a50bb27edba8176e69f87d0ec5cf386ed6867d4bf6 Dec 03 09:06:48 crc kubenswrapper[4573]: I1203 09:06:48.924905 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" event={"ID":"609d7944-6cf6-47c4-a1f2-2437c57703d2","Type":"ContainerStarted","Data":"93e968bb43e51dce488aa5c8bb123a2864d3c2f8b0271b8b49f388400d7ae6eb"} Dec 03 09:06:48 crc kubenswrapper[4573]: I1203 09:06:48.925550 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" event={"ID":"609d7944-6cf6-47c4-a1f2-2437c57703d2","Type":"ContainerStarted","Data":"4f484ac1378459a570a1e2a50bb27edba8176e69f87d0ec5cf386ed6867d4bf6"} Dec 03 09:06:48 crc kubenswrapper[4573]: I1203 09:06:48.958984 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" podStartSLOduration=1.7410939010000002 podStartE2EDuration="1.958960189s" podCreationTimestamp="2025-12-03 09:06:47 +0000 UTC" firstStartedPulling="2025-12-03 09:06:48.027317192 +0000 UTC m=+1728.595696451" lastFinishedPulling="2025-12-03 09:06:48.24518348 +0000 UTC m=+1728.813562739" observedRunningTime="2025-12-03 09:06:48.944842341 +0000 UTC m=+1729.513221640" watchObservedRunningTime="2025-12-03 09:06:48.958960189 +0000 UTC m=+1729.527339448" Dec 03 09:06:50 crc kubenswrapper[4573]: I1203 09:06:50.038867 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:06:50 crc kubenswrapper[4573]: E1203 09:06:50.039685 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:06:51 crc kubenswrapper[4573]: I1203 09:06:51.953879 4573 generic.go:334] "Generic (PLEG): container finished" podID="609d7944-6cf6-47c4-a1f2-2437c57703d2" containerID="93e968bb43e51dce488aa5c8bb123a2864d3c2f8b0271b8b49f388400d7ae6eb" exitCode=0 Dec 03 09:06:51 crc kubenswrapper[4573]: I1203 09:06:51.953928 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" event={"ID":"609d7944-6cf6-47c4-a1f2-2437c57703d2","Type":"ContainerDied","Data":"93e968bb43e51dce488aa5c8bb123a2864d3c2f8b0271b8b49f388400d7ae6eb"} Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.444042 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.528180 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz6rk\" (UniqueName: \"kubernetes.io/projected/609d7944-6cf6-47c4-a1f2-2437c57703d2-kube-api-access-gz6rk\") pod \"609d7944-6cf6-47c4-a1f2-2437c57703d2\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.528235 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-inventory\") pod \"609d7944-6cf6-47c4-a1f2-2437c57703d2\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.528289 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-ssh-key\") pod \"609d7944-6cf6-47c4-a1f2-2437c57703d2\" (UID: \"609d7944-6cf6-47c4-a1f2-2437c57703d2\") " Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.550146 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/609d7944-6cf6-47c4-a1f2-2437c57703d2-kube-api-access-gz6rk" (OuterVolumeSpecName: "kube-api-access-gz6rk") pod "609d7944-6cf6-47c4-a1f2-2437c57703d2" (UID: "609d7944-6cf6-47c4-a1f2-2437c57703d2"). InnerVolumeSpecName "kube-api-access-gz6rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.562551 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-inventory" (OuterVolumeSpecName: "inventory") pod "609d7944-6cf6-47c4-a1f2-2437c57703d2" (UID: "609d7944-6cf6-47c4-a1f2-2437c57703d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.568015 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "609d7944-6cf6-47c4-a1f2-2437c57703d2" (UID: "609d7944-6cf6-47c4-a1f2-2437c57703d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.631235 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz6rk\" (UniqueName: \"kubernetes.io/projected/609d7944-6cf6-47c4-a1f2-2437c57703d2-kube-api-access-gz6rk\") on node \"crc\" DevicePath \"\"" Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.631277 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.631287 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/609d7944-6cf6-47c4-a1f2-2437c57703d2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.978300 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" event={"ID":"609d7944-6cf6-47c4-a1f2-2437c57703d2","Type":"ContainerDied","Data":"4f484ac1378459a570a1e2a50bb27edba8176e69f87d0ec5cf386ed6867d4bf6"} Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.978346 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f484ac1378459a570a1e2a50bb27edba8176e69f87d0ec5cf386ed6867d4bf6" Dec 03 09:06:53 crc kubenswrapper[4573]: I1203 09:06:53.978386 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-lbctc" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.092309 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r"] Dec 03 09:06:54 crc kubenswrapper[4573]: E1203 09:06:54.092787 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="609d7944-6cf6-47c4-a1f2-2437c57703d2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.092809 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="609d7944-6cf6-47c4-a1f2-2437c57703d2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.092977 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="609d7944-6cf6-47c4-a1f2-2437c57703d2" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.093757 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.108487 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.108770 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.108956 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.109169 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.114520 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r"] Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.143428 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.143520 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4nkz\" (UniqueName: \"kubernetes.io/projected/3dc88703-275c-419a-961e-9c034464b6cb-kube-api-access-n4nkz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.143693 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.143928 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.246206 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.246297 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4nkz\" (UniqueName: \"kubernetes.io/projected/3dc88703-275c-419a-961e-9c034464b6cb-kube-api-access-n4nkz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.246421 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.246513 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.251808 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.257464 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.257590 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.268356 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4nkz\" (UniqueName: \"kubernetes.io/projected/3dc88703-275c-419a-961e-9c034464b6cb-kube-api-access-n4nkz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:54 crc kubenswrapper[4573]: I1203 09:06:54.428240 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:06:55 crc kubenswrapper[4573]: I1203 09:06:55.009791 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r"] Dec 03 09:06:56 crc kubenswrapper[4573]: I1203 09:06:56.004436 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" event={"ID":"3dc88703-275c-419a-961e-9c034464b6cb","Type":"ContainerStarted","Data":"ce399b50f2ae4b581d8454247b6cbead598dcf42a2ec70e5a2789dccd203111c"} Dec 03 09:06:56 crc kubenswrapper[4573]: I1203 09:06:56.004487 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" event={"ID":"3dc88703-275c-419a-961e-9c034464b6cb","Type":"ContainerStarted","Data":"3534364eb03caf99a980d427da7a3d74c288ae588aefe61a74f221c7696a3b3c"} Dec 03 09:06:56 crc kubenswrapper[4573]: I1203 09:06:56.020841 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" podStartSLOduration=1.842402707 podStartE2EDuration="2.02081799s" podCreationTimestamp="2025-12-03 09:06:54 +0000 UTC" firstStartedPulling="2025-12-03 09:06:55.020728912 +0000 UTC m=+1735.589108171" lastFinishedPulling="2025-12-03 09:06:55.199144195 +0000 UTC m=+1735.767523454" observedRunningTime="2025-12-03 09:06:56.019706719 +0000 UTC m=+1736.588085998" watchObservedRunningTime="2025-12-03 09:06:56.02081799 +0000 UTC m=+1736.589197259" Dec 03 09:07:02 crc kubenswrapper[4573]: I1203 09:07:02.030789 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:07:02 crc kubenswrapper[4573]: E1203 09:07:02.031849 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:07:15 crc kubenswrapper[4573]: I1203 09:07:15.030878 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:07:15 crc kubenswrapper[4573]: E1203 09:07:15.031993 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:07:26 crc kubenswrapper[4573]: I1203 09:07:26.031716 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:07:26 crc kubenswrapper[4573]: E1203 09:07:26.032696 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:07:41 crc kubenswrapper[4573]: I1203 09:07:41.030986 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:07:41 crc kubenswrapper[4573]: E1203 09:07:41.031719 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:07:54 crc kubenswrapper[4573]: I1203 09:07:54.031298 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:07:54 crc kubenswrapper[4573]: E1203 09:07:54.033362 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:08:05 crc kubenswrapper[4573]: I1203 09:08:05.031431 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:08:05 crc kubenswrapper[4573]: E1203 09:08:05.032871 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:08:19 crc kubenswrapper[4573]: I1203 09:08:19.031615 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:08:19 crc kubenswrapper[4573]: E1203 09:08:19.032989 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:08:34 crc kubenswrapper[4573]: I1203 09:08:34.030447 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:08:34 crc kubenswrapper[4573]: E1203 09:08:34.033224 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:08:49 crc kubenswrapper[4573]: I1203 09:08:49.031308 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:08:49 crc kubenswrapper[4573]: E1203 09:08:49.034565 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:09:04 crc kubenswrapper[4573]: I1203 09:09:04.043630 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:09:04 crc kubenswrapper[4573]: E1203 09:09:04.047477 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:09:08 crc kubenswrapper[4573]: I1203 09:09:08.076338 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-74eb-account-create-update-hhzh4"] Dec 03 09:09:08 crc kubenswrapper[4573]: I1203 09:09:08.096670 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-nfl5h"] Dec 03 09:09:08 crc kubenswrapper[4573]: I1203 09:09:08.107455 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-nfl5h"] Dec 03 09:09:08 crc kubenswrapper[4573]: I1203 09:09:08.120303 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-74eb-account-create-update-hhzh4"] Dec 03 09:09:09 crc kubenswrapper[4573]: I1203 09:09:09.048249 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-e92a-account-create-update-6bjzf"] Dec 03 09:09:09 crc kubenswrapper[4573]: I1203 09:09:09.058527 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-e92a-account-create-update-6bjzf"] Dec 03 09:09:10 crc kubenswrapper[4573]: I1203 09:09:10.050495 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81d9684c-2d7b-49df-a5c6-a56656a84874" path="/var/lib/kubelet/pods/81d9684c-2d7b-49df-a5c6-a56656a84874/volumes" Dec 03 09:09:10 crc kubenswrapper[4573]: I1203 09:09:10.051418 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd12c3b6-45e8-44dd-b28f-b7c828cb5690" path="/var/lib/kubelet/pods/dd12c3b6-45e8-44dd-b28f-b7c828cb5690/volumes" Dec 03 09:09:10 crc kubenswrapper[4573]: I1203 09:09:10.052762 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78a67e6-fbd8-4c2b-bc2d-148de6b27c20" path="/var/lib/kubelet/pods/e78a67e6-fbd8-4c2b-bc2d-148de6b27c20/volumes" Dec 03 09:09:10 crc kubenswrapper[4573]: I1203 09:09:10.072146 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-6p2pk"] Dec 03 09:09:10 crc kubenswrapper[4573]: I1203 09:09:10.085516 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-6p2pk"] Dec 03 09:09:12 crc kubenswrapper[4573]: I1203 09:09:12.084577 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0d549a7-b237-46de-8dde-d7aba95cfe65" path="/var/lib/kubelet/pods/f0d549a7-b237-46de-8dde-d7aba95cfe65/volumes" Dec 03 09:09:12 crc kubenswrapper[4573]: I1203 09:09:12.085799 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-xd2w2"] Dec 03 09:09:12 crc kubenswrapper[4573]: I1203 09:09:12.093882 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-3c12-account-create-update-ntcsm"] Dec 03 09:09:12 crc kubenswrapper[4573]: I1203 09:09:12.116478 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-xd2w2"] Dec 03 09:09:12 crc kubenswrapper[4573]: I1203 09:09:12.132481 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-3c12-account-create-update-ntcsm"] Dec 03 09:09:14 crc kubenswrapper[4573]: I1203 09:09:14.049981 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d62ff626-7df4-4756-ab26-a2d74f71bb57" path="/var/lib/kubelet/pods/d62ff626-7df4-4756-ab26-a2d74f71bb57/volumes" Dec 03 09:09:14 crc kubenswrapper[4573]: I1203 09:09:14.051300 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffbef148-b74e-4d49-ab26-1c1e9808074f" path="/var/lib/kubelet/pods/ffbef148-b74e-4d49-ab26-1c1e9808074f/volumes" Dec 03 09:09:19 crc kubenswrapper[4573]: I1203 09:09:19.031160 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:09:19 crc kubenswrapper[4573]: E1203 09:09:19.032163 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:09:30 crc kubenswrapper[4573]: I1203 09:09:30.043718 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:09:30 crc kubenswrapper[4573]: E1203 09:09:30.044916 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:09:43 crc kubenswrapper[4573]: I1203 09:09:43.030715 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:09:43 crc kubenswrapper[4573]: E1203 09:09:43.031486 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:09:45 crc kubenswrapper[4573]: I1203 09:09:45.047828 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b911-account-create-update-54pl8"] Dec 03 09:09:45 crc kubenswrapper[4573]: I1203 09:09:45.063092 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-hfvkh"] Dec 03 09:09:45 crc kubenswrapper[4573]: I1203 09:09:45.074428 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-v5rph"] Dec 03 09:09:45 crc kubenswrapper[4573]: I1203 09:09:45.084470 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-v5rph"] Dec 03 09:09:45 crc kubenswrapper[4573]: I1203 09:09:45.092971 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b911-account-create-update-54pl8"] Dec 03 09:09:45 crc kubenswrapper[4573]: I1203 09:09:45.101384 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-hfvkh"] Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.042340 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ff8212d-0f81-4ecd-88ee-cfcc421a1747" path="/var/lib/kubelet/pods/8ff8212d-0f81-4ecd-88ee-cfcc421a1747/volumes" Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.043605 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e" path="/var/lib/kubelet/pods/9fb93be9-0ba4-4e21-8a85-7aca6c2dbf7e/volumes" Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.044389 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7797a3f-2e04-44d9-a0f9-9958389b9810" path="/var/lib/kubelet/pods/f7797a3f-2e04-44d9-a0f9-9958389b9810/volumes" Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.050362 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-019c-account-create-update-kplbv"] Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.058221 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-74c1-account-create-update-2ssqv"] Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.069063 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-019c-account-create-update-kplbv"] Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.079084 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-hklh5"] Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.089340 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-74c1-account-create-update-2ssqv"] Dec 03 09:09:46 crc kubenswrapper[4573]: I1203 09:09:46.099782 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-hklh5"] Dec 03 09:09:48 crc kubenswrapper[4573]: I1203 09:09:48.040790 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38e08930-afea-488b-b008-f9d92b8671de" path="/var/lib/kubelet/pods/38e08930-afea-488b-b008-f9d92b8671de/volumes" Dec 03 09:09:48 crc kubenswrapper[4573]: I1203 09:09:48.041779 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781" path="/var/lib/kubelet/pods/62f89ce0-bc6c-4c50-adcf-3cfbc8c5c781/volumes" Dec 03 09:09:48 crc kubenswrapper[4573]: I1203 09:09:48.042564 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc73654f-5d5d-4b47-ab85-fd28bd21bed2" path="/var/lib/kubelet/pods/bc73654f-5d5d-4b47-ab85-fd28bd21bed2/volumes" Dec 03 09:09:55 crc kubenswrapper[4573]: I1203 09:09:55.032661 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:09:55 crc kubenswrapper[4573]: E1203 09:09:55.033854 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:09:58 crc kubenswrapper[4573]: I1203 09:09:58.042887 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-m8n6d"] Dec 03 09:09:58 crc kubenswrapper[4573]: I1203 09:09:58.044463 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-m8n6d"] Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.049367 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ca3bd5-ac7a-40c1-b9b9-394ee79f5654" path="/var/lib/kubelet/pods/09ca3bd5-ac7a-40c1-b9b9-394ee79f5654/volumes" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.631855 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gng9s"] Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.634470 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.664108 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gng9s"] Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.818558 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7wr8\" (UniqueName: \"kubernetes.io/projected/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-kube-api-access-x7wr8\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.818652 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-utilities\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.819874 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-catalog-content\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.923420 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7wr8\" (UniqueName: \"kubernetes.io/projected/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-kube-api-access-x7wr8\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.923536 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-utilities\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.923577 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-catalog-content\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.924658 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-catalog-content\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.925575 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-utilities\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.949746 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7wr8\" (UniqueName: \"kubernetes.io/projected/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-kube-api-access-x7wr8\") pod \"redhat-marketplace-gng9s\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:00 crc kubenswrapper[4573]: I1203 09:10:00.960143 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:01 crc kubenswrapper[4573]: I1203 09:10:01.658889 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gng9s"] Dec 03 09:10:02 crc kubenswrapper[4573]: I1203 09:10:02.155664 4573 generic.go:334] "Generic (PLEG): container finished" podID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerID="58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9" exitCode=0 Dec 03 09:10:02 crc kubenswrapper[4573]: I1203 09:10:02.156222 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gng9s" event={"ID":"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff","Type":"ContainerDied","Data":"58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9"} Dec 03 09:10:02 crc kubenswrapper[4573]: I1203 09:10:02.156266 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gng9s" event={"ID":"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff","Type":"ContainerStarted","Data":"d0c727ffd329bc6ab1217e77379026c32b72e64f73f8151bde2851279e2c2e84"} Dec 03 09:10:02 crc kubenswrapper[4573]: I1203 09:10:02.160526 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.049448 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bnzb5"] Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.053952 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.073677 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnzb5"] Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.180892 4573 generic.go:334] "Generic (PLEG): container finished" podID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerID="86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65" exitCode=0 Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.180959 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gng9s" event={"ID":"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff","Type":"ContainerDied","Data":"86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65"} Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.209953 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-catalog-content\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.210136 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vnnw\" (UniqueName: \"kubernetes.io/projected/751ca9a5-70b6-4548-93f8-031c471dc28e-kube-api-access-5vnnw\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.210393 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-utilities\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.311967 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-utilities\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.312417 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-catalog-content\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.312602 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vnnw\" (UniqueName: \"kubernetes.io/projected/751ca9a5-70b6-4548-93f8-031c471dc28e-kube-api-access-5vnnw\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.312602 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-utilities\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.312915 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-catalog-content\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.353430 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vnnw\" (UniqueName: \"kubernetes.io/projected/751ca9a5-70b6-4548-93f8-031c471dc28e-kube-api-access-5vnnw\") pod \"community-operators-bnzb5\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:04 crc kubenswrapper[4573]: I1203 09:10:04.391815 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.193088 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gng9s" event={"ID":"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff","Type":"ContainerStarted","Data":"1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b"} Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.219625 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gng9s" podStartSLOduration=2.629426525 podStartE2EDuration="5.219589164s" podCreationTimestamp="2025-12-03 09:10:00 +0000 UTC" firstStartedPulling="2025-12-03 09:10:02.16027943 +0000 UTC m=+1922.728658689" lastFinishedPulling="2025-12-03 09:10:04.750442069 +0000 UTC m=+1925.318821328" observedRunningTime="2025-12-03 09:10:05.216291894 +0000 UTC m=+1925.784671163" watchObservedRunningTime="2025-12-03 09:10:05.219589164 +0000 UTC m=+1925.787968423" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.246880 4573 scope.go:117] "RemoveContainer" containerID="1018de56d10637ce91f2fd6f76a572850cd63f4ac84753bca8cc1328b10d4cf7" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.286842 4573 scope.go:117] "RemoveContainer" containerID="ade849468c4023a969b366cfc153daa2689586665d15505ccda22d0c6012dfcb" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.295431 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnzb5"] Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.369234 4573 scope.go:117] "RemoveContainer" containerID="e2f567c239929f4b9f39cffda960b440fefc7418e809061dabfe475e2ab84f84" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.443022 4573 scope.go:117] "RemoveContainer" containerID="ffc97e4137bae2488e3f6389c3fb5e75e5948c598147f5d1e8772f55cb0acb3d" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.541493 4573 scope.go:117] "RemoveContainer" containerID="a8c37657b2f5e149a0ffbf74f589cc9661b59ae72e309ebeecfdf88a8046472a" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.621823 4573 scope.go:117] "RemoveContainer" containerID="07f699b6a8d9d47e724de128143d562bd9b9db501e43b50f272b4a4b256ef4db" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.661405 4573 scope.go:117] "RemoveContainer" containerID="c19b4d5e1262e2b1bb5fd37ae2944d1b0fb9eeee59d7cf0ec7e54a49bb92c0c6" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.763793 4573 scope.go:117] "RemoveContainer" containerID="f50555e5e1b886b27f9be5f4290762ef64b328d70615ead4dc38bf3a0cf88cd2" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.829062 4573 scope.go:117] "RemoveContainer" containerID="ca2e5f9be83477a0de46d55c12f120cdb471db124cc18e1b801aa299dd9e0563" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.883260 4573 scope.go:117] "RemoveContainer" containerID="676084a6714120ab873f67d0935d1df84f7b21662a8b80eb74d9ce58f07d9d44" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.936988 4573 scope.go:117] "RemoveContainer" containerID="7f083f03abb690a4d24279e62b5db12da2982605d805b23ec683050373dbb0e5" Dec 03 09:10:05 crc kubenswrapper[4573]: I1203 09:10:05.987466 4573 scope.go:117] "RemoveContainer" containerID="a8254e75bc483a40ed1b50671feca39018863f80357e2b79c542ea378cfc4d46" Dec 03 09:10:06 crc kubenswrapper[4573]: I1203 09:10:06.052597 4573 scope.go:117] "RemoveContainer" containerID="52472c809a49a3475105ccfba04e5a080eb833b200db80ba885d49dfed5e4281" Dec 03 09:10:06 crc kubenswrapper[4573]: I1203 09:10:06.213435 4573 generic.go:334] "Generic (PLEG): container finished" podID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerID="50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff" exitCode=0 Dec 03 09:10:06 crc kubenswrapper[4573]: I1203 09:10:06.213508 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnzb5" event={"ID":"751ca9a5-70b6-4548-93f8-031c471dc28e","Type":"ContainerDied","Data":"50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff"} Dec 03 09:10:06 crc kubenswrapper[4573]: I1203 09:10:06.213532 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnzb5" event={"ID":"751ca9a5-70b6-4548-93f8-031c471dc28e","Type":"ContainerStarted","Data":"35402bdc87f55f2e66e3777d8bb59b487986faf87e192b8d00b700e056578c5c"} Dec 03 09:10:07 crc kubenswrapper[4573]: I1203 09:10:07.030872 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:10:07 crc kubenswrapper[4573]: E1203 09:10:07.031616 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:10:07 crc kubenswrapper[4573]: I1203 09:10:07.288440 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnzb5" event={"ID":"751ca9a5-70b6-4548-93f8-031c471dc28e","Type":"ContainerStarted","Data":"3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0"} Dec 03 09:10:10 crc kubenswrapper[4573]: I1203 09:10:10.961613 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:10 crc kubenswrapper[4573]: I1203 09:10:10.961953 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:11 crc kubenswrapper[4573]: I1203 09:10:11.038608 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:11 crc kubenswrapper[4573]: I1203 09:10:11.335579 4573 generic.go:334] "Generic (PLEG): container finished" podID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerID="3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0" exitCode=0 Dec 03 09:10:11 crc kubenswrapper[4573]: I1203 09:10:11.335648 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnzb5" event={"ID":"751ca9a5-70b6-4548-93f8-031c471dc28e","Type":"ContainerDied","Data":"3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0"} Dec 03 09:10:11 crc kubenswrapper[4573]: I1203 09:10:11.409437 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:12 crc kubenswrapper[4573]: I1203 09:10:12.359816 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnzb5" event={"ID":"751ca9a5-70b6-4548-93f8-031c471dc28e","Type":"ContainerStarted","Data":"7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26"} Dec 03 09:10:13 crc kubenswrapper[4573]: I1203 09:10:13.304560 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gng9s"] Dec 03 09:10:13 crc kubenswrapper[4573]: I1203 09:10:13.371832 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gng9s" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerName="registry-server" containerID="cri-o://1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b" gracePeriod=2 Dec 03 09:10:13 crc kubenswrapper[4573]: I1203 09:10:13.414721 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bnzb5" podStartSLOduration=3.832779242 podStartE2EDuration="9.41469522s" podCreationTimestamp="2025-12-03 09:10:04 +0000 UTC" firstStartedPulling="2025-12-03 09:10:06.237280482 +0000 UTC m=+1926.805659741" lastFinishedPulling="2025-12-03 09:10:11.81919646 +0000 UTC m=+1932.387575719" observedRunningTime="2025-12-03 09:10:13.413187089 +0000 UTC m=+1933.981566358" watchObservedRunningTime="2025-12-03 09:10:13.41469522 +0000 UTC m=+1933.983074479" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.211505 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.364564 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-utilities\") pod \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.364625 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7wr8\" (UniqueName: \"kubernetes.io/projected/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-kube-api-access-x7wr8\") pod \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.364667 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-catalog-content\") pod \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\" (UID: \"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff\") " Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.365957 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-utilities" (OuterVolumeSpecName: "utilities") pod "fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" (UID: "fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.383519 4573 generic.go:334] "Generic (PLEG): container finished" podID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerID="1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b" exitCode=0 Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.383562 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gng9s" event={"ID":"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff","Type":"ContainerDied","Data":"1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b"} Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.383578 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gng9s" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.383691 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gng9s" event={"ID":"fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff","Type":"ContainerDied","Data":"d0c727ffd329bc6ab1217e77379026c32b72e64f73f8151bde2851279e2c2e84"} Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.383784 4573 scope.go:117] "RemoveContainer" containerID="1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.386235 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-kube-api-access-x7wr8" (OuterVolumeSpecName: "kube-api-access-x7wr8") pod "fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" (UID: "fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff"). InnerVolumeSpecName "kube-api-access-x7wr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.391690 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" (UID: "fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.392035 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.395074 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.436472 4573 scope.go:117] "RemoveContainer" containerID="86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.469251 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.469299 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7wr8\" (UniqueName: \"kubernetes.io/projected/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-kube-api-access-x7wr8\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.469312 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.482067 4573 scope.go:117] "RemoveContainer" containerID="58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.549000 4573 scope.go:117] "RemoveContainer" containerID="1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b" Dec 03 09:10:14 crc kubenswrapper[4573]: E1203 09:10:14.549683 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b\": container with ID starting with 1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b not found: ID does not exist" containerID="1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.549716 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b"} err="failed to get container status \"1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b\": rpc error: code = NotFound desc = could not find container \"1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b\": container with ID starting with 1eaafe798693de43c3cfc8bcb4ed200f0a843a233901ee4c222e1648b5dd684b not found: ID does not exist" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.549742 4573 scope.go:117] "RemoveContainer" containerID="86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65" Dec 03 09:10:14 crc kubenswrapper[4573]: E1203 09:10:14.550002 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65\": container with ID starting with 86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65 not found: ID does not exist" containerID="86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.551520 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65"} err="failed to get container status \"86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65\": rpc error: code = NotFound desc = could not find container \"86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65\": container with ID starting with 86c8932b2177999b18bcdb39f871753000ca9b227ccf4c2713c028a68ed7fb65 not found: ID does not exist" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.551550 4573 scope.go:117] "RemoveContainer" containerID="58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9" Dec 03 09:10:14 crc kubenswrapper[4573]: E1203 09:10:14.552134 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9\": container with ID starting with 58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9 not found: ID does not exist" containerID="58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.552161 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9"} err="failed to get container status \"58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9\": rpc error: code = NotFound desc = could not find container \"58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9\": container with ID starting with 58d414f5de7ec025c67271c2e906ad8cebdd6bb04a29282979060db3ca90c8a9 not found: ID does not exist" Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.746124 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gng9s"] Dec 03 09:10:14 crc kubenswrapper[4573]: I1203 09:10:14.766469 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gng9s"] Dec 03 09:10:15 crc kubenswrapper[4573]: I1203 09:10:15.541500 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-bnzb5" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="registry-server" probeResult="failure" output=< Dec 03 09:10:15 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:10:15 crc kubenswrapper[4573]: > Dec 03 09:10:16 crc kubenswrapper[4573]: I1203 09:10:16.041879 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" path="/var/lib/kubelet/pods/fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff/volumes" Dec 03 09:10:20 crc kubenswrapper[4573]: I1203 09:10:20.037326 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:10:20 crc kubenswrapper[4573]: E1203 09:10:20.038170 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:10:24 crc kubenswrapper[4573]: I1203 09:10:24.456971 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:24 crc kubenswrapper[4573]: I1203 09:10:24.520671 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:24 crc kubenswrapper[4573]: I1203 09:10:24.695294 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bnzb5"] Dec 03 09:10:25 crc kubenswrapper[4573]: I1203 09:10:25.501192 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bnzb5" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="registry-server" containerID="cri-o://7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26" gracePeriod=2 Dec 03 09:10:25 crc kubenswrapper[4573]: I1203 09:10:25.998159 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.197082 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-utilities\") pod \"751ca9a5-70b6-4548-93f8-031c471dc28e\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.197141 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vnnw\" (UniqueName: \"kubernetes.io/projected/751ca9a5-70b6-4548-93f8-031c471dc28e-kube-api-access-5vnnw\") pod \"751ca9a5-70b6-4548-93f8-031c471dc28e\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.197250 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-catalog-content\") pod \"751ca9a5-70b6-4548-93f8-031c471dc28e\" (UID: \"751ca9a5-70b6-4548-93f8-031c471dc28e\") " Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.198303 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-utilities" (OuterVolumeSpecName: "utilities") pod "751ca9a5-70b6-4548-93f8-031c471dc28e" (UID: "751ca9a5-70b6-4548-93f8-031c471dc28e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.228629 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751ca9a5-70b6-4548-93f8-031c471dc28e-kube-api-access-5vnnw" (OuterVolumeSpecName: "kube-api-access-5vnnw") pod "751ca9a5-70b6-4548-93f8-031c471dc28e" (UID: "751ca9a5-70b6-4548-93f8-031c471dc28e"). InnerVolumeSpecName "kube-api-access-5vnnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.275066 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "751ca9a5-70b6-4548-93f8-031c471dc28e" (UID: "751ca9a5-70b6-4548-93f8-031c471dc28e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.299947 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.299999 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vnnw\" (UniqueName: \"kubernetes.io/projected/751ca9a5-70b6-4548-93f8-031c471dc28e-kube-api-access-5vnnw\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.300013 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/751ca9a5-70b6-4548-93f8-031c471dc28e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.512391 4573 generic.go:334] "Generic (PLEG): container finished" podID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerID="7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26" exitCode=0 Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.512441 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnzb5" event={"ID":"751ca9a5-70b6-4548-93f8-031c471dc28e","Type":"ContainerDied","Data":"7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26"} Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.512479 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnzb5" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.512509 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnzb5" event={"ID":"751ca9a5-70b6-4548-93f8-031c471dc28e","Type":"ContainerDied","Data":"35402bdc87f55f2e66e3777d8bb59b487986faf87e192b8d00b700e056578c5c"} Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.512535 4573 scope.go:117] "RemoveContainer" containerID="7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.540110 4573 scope.go:117] "RemoveContainer" containerID="3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.570630 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bnzb5"] Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.575826 4573 scope.go:117] "RemoveContainer" containerID="50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.586688 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bnzb5"] Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.619390 4573 scope.go:117] "RemoveContainer" containerID="7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26" Dec 03 09:10:26 crc kubenswrapper[4573]: E1203 09:10:26.620070 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26\": container with ID starting with 7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26 not found: ID does not exist" containerID="7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.620131 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26"} err="failed to get container status \"7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26\": rpc error: code = NotFound desc = could not find container \"7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26\": container with ID starting with 7f65569fbc722d58a983e40b831c963241e947918035a945d558ebf0364b5f26 not found: ID does not exist" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.620168 4573 scope.go:117] "RemoveContainer" containerID="3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0" Dec 03 09:10:26 crc kubenswrapper[4573]: E1203 09:10:26.620583 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0\": container with ID starting with 3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0 not found: ID does not exist" containerID="3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.620666 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0"} err="failed to get container status \"3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0\": rpc error: code = NotFound desc = could not find container \"3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0\": container with ID starting with 3340c7b18ea64c95c5ac01957c6695a6194ec50e77233d637c24f3064cdccdd0 not found: ID does not exist" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.620690 4573 scope.go:117] "RemoveContainer" containerID="50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff" Dec 03 09:10:26 crc kubenswrapper[4573]: E1203 09:10:26.620966 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff\": container with ID starting with 50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff not found: ID does not exist" containerID="50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff" Dec 03 09:10:26 crc kubenswrapper[4573]: I1203 09:10:26.620995 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff"} err="failed to get container status \"50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff\": rpc error: code = NotFound desc = could not find container \"50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff\": container with ID starting with 50b9678429e438d11640b7606c2a856f34d423157e72dcfd76de9e29e975a1ff not found: ID does not exist" Dec 03 09:10:28 crc kubenswrapper[4573]: I1203 09:10:28.050404 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" path="/var/lib/kubelet/pods/751ca9a5-70b6-4548-93f8-031c471dc28e/volumes" Dec 03 09:10:31 crc kubenswrapper[4573]: I1203 09:10:31.053101 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-8t69f"] Dec 03 09:10:31 crc kubenswrapper[4573]: I1203 09:10:31.061543 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-8t69f"] Dec 03 09:10:32 crc kubenswrapper[4573]: I1203 09:10:32.031314 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:10:32 crc kubenswrapper[4573]: E1203 09:10:32.031847 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:10:32 crc kubenswrapper[4573]: I1203 09:10:32.048363 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de65a9c9-bef6-42e3-b256-f94d4af37c96" path="/var/lib/kubelet/pods/de65a9c9-bef6-42e3-b256-f94d4af37c96/volumes" Dec 03 09:10:45 crc kubenswrapper[4573]: I1203 09:10:45.030230 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:10:45 crc kubenswrapper[4573]: E1203 09:10:45.031132 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:10:49 crc kubenswrapper[4573]: I1203 09:10:49.767253 4573 generic.go:334] "Generic (PLEG): container finished" podID="3dc88703-275c-419a-961e-9c034464b6cb" containerID="ce399b50f2ae4b581d8454247b6cbead598dcf42a2ec70e5a2789dccd203111c" exitCode=0 Dec 03 09:10:49 crc kubenswrapper[4573]: I1203 09:10:49.767663 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" event={"ID":"3dc88703-275c-419a-961e-9c034464b6cb","Type":"ContainerDied","Data":"ce399b50f2ae4b581d8454247b6cbead598dcf42a2ec70e5a2789dccd203111c"} Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.274974 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.329182 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4nkz\" (UniqueName: \"kubernetes.io/projected/3dc88703-275c-419a-961e-9c034464b6cb-kube-api-access-n4nkz\") pod \"3dc88703-275c-419a-961e-9c034464b6cb\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.329702 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-ssh-key\") pod \"3dc88703-275c-419a-961e-9c034464b6cb\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.330024 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-inventory\") pod \"3dc88703-275c-419a-961e-9c034464b6cb\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.330346 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-bootstrap-combined-ca-bundle\") pod \"3dc88703-275c-419a-961e-9c034464b6cb\" (UID: \"3dc88703-275c-419a-961e-9c034464b6cb\") " Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.355269 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3dc88703-275c-419a-961e-9c034464b6cb" (UID: "3dc88703-275c-419a-961e-9c034464b6cb"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.363562 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc88703-275c-419a-961e-9c034464b6cb-kube-api-access-n4nkz" (OuterVolumeSpecName: "kube-api-access-n4nkz") pod "3dc88703-275c-419a-961e-9c034464b6cb" (UID: "3dc88703-275c-419a-961e-9c034464b6cb"). InnerVolumeSpecName "kube-api-access-n4nkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.371037 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3dc88703-275c-419a-961e-9c034464b6cb" (UID: "3dc88703-275c-419a-961e-9c034464b6cb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.391079 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-inventory" (OuterVolumeSpecName: "inventory") pod "3dc88703-275c-419a-961e-9c034464b6cb" (UID: "3dc88703-275c-419a-961e-9c034464b6cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.431718 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.431767 4573 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.431782 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4nkz\" (UniqueName: \"kubernetes.io/projected/3dc88703-275c-419a-961e-9c034464b6cb-kube-api-access-n4nkz\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.431794 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3dc88703-275c-419a-961e-9c034464b6cb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.792160 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" event={"ID":"3dc88703-275c-419a-961e-9c034464b6cb","Type":"ContainerDied","Data":"3534364eb03caf99a980d427da7a3d74c288ae588aefe61a74f221c7696a3b3c"} Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.792211 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3534364eb03caf99a980d427da7a3d74c288ae588aefe61a74f221c7696a3b3c" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.792322 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.932989 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67"] Dec 03 09:10:51 crc kubenswrapper[4573]: E1203 09:10:51.935734 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="registry-server" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.935839 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="registry-server" Dec 03 09:10:51 crc kubenswrapper[4573]: E1203 09:10:51.935905 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerName="extract-content" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.935957 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerName="extract-content" Dec 03 09:10:51 crc kubenswrapper[4573]: E1203 09:10:51.936024 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerName="registry-server" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.936103 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerName="registry-server" Dec 03 09:10:51 crc kubenswrapper[4573]: E1203 09:10:51.936199 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc88703-275c-419a-961e-9c034464b6cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.936258 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc88703-275c-419a-961e-9c034464b6cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 09:10:51 crc kubenswrapper[4573]: E1203 09:10:51.936320 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="extract-content" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.936371 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="extract-content" Dec 03 09:10:51 crc kubenswrapper[4573]: E1203 09:10:51.936434 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerName="extract-utilities" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.936486 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerName="extract-utilities" Dec 03 09:10:51 crc kubenswrapper[4573]: E1203 09:10:51.936567 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="extract-utilities" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.936653 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="extract-utilities" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.936931 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="751ca9a5-70b6-4548-93f8-031c471dc28e" containerName="registry-server" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.937092 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc88703-275c-419a-961e-9c034464b6cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.937198 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa7d9eb8-0ce8-4bcc-8ead-a728e329a2ff" containerName="registry-server" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.938348 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.941565 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.943708 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.943795 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.944180 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.943984 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.944434 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.944610 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld725\" (UniqueName: \"kubernetes.io/projected/960face3-051e-4d24-b7ae-8e8b0f13d14e-kube-api-access-ld725\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:51 crc kubenswrapper[4573]: I1203 09:10:51.954964 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67"] Dec 03 09:10:52 crc kubenswrapper[4573]: I1203 09:10:52.046471 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:52 crc kubenswrapper[4573]: I1203 09:10:52.047210 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:52 crc kubenswrapper[4573]: I1203 09:10:52.047274 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld725\" (UniqueName: \"kubernetes.io/projected/960face3-051e-4d24-b7ae-8e8b0f13d14e-kube-api-access-ld725\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:52 crc kubenswrapper[4573]: I1203 09:10:52.051952 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:52 crc kubenswrapper[4573]: I1203 09:10:52.052679 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:52 crc kubenswrapper[4573]: I1203 09:10:52.068170 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld725\" (UniqueName: \"kubernetes.io/projected/960face3-051e-4d24-b7ae-8e8b0f13d14e-kube-api-access-ld725\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6kp67\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:52 crc kubenswrapper[4573]: I1203 09:10:52.257142 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:10:52 crc kubenswrapper[4573]: I1203 09:10:52.829799 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67"] Dec 03 09:10:53 crc kubenswrapper[4573]: I1203 09:10:53.075420 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-srtdq"] Dec 03 09:10:53 crc kubenswrapper[4573]: I1203 09:10:53.140904 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-srtdq"] Dec 03 09:10:53 crc kubenswrapper[4573]: I1203 09:10:53.816799 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" event={"ID":"960face3-051e-4d24-b7ae-8e8b0f13d14e","Type":"ContainerStarted","Data":"297f3294e1cc02316d1b7b52b73510727c2e2ad25d5ecda1d65caddeb2a63338"} Dec 03 09:10:53 crc kubenswrapper[4573]: I1203 09:10:53.817605 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" event={"ID":"960face3-051e-4d24-b7ae-8e8b0f13d14e","Type":"ContainerStarted","Data":"02665d640b27316fd195ded4c56a9cc7e5bdfd19ef4ea1c29bfeda0696c065fe"} Dec 03 09:10:53 crc kubenswrapper[4573]: I1203 09:10:53.848726 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" podStartSLOduration=2.638709015 podStartE2EDuration="2.848699967s" podCreationTimestamp="2025-12-03 09:10:51 +0000 UTC" firstStartedPulling="2025-12-03 09:10:52.828301316 +0000 UTC m=+1973.396680575" lastFinishedPulling="2025-12-03 09:10:53.038292268 +0000 UTC m=+1973.606671527" observedRunningTime="2025-12-03 09:10:53.838590422 +0000 UTC m=+1974.406969701" watchObservedRunningTime="2025-12-03 09:10:53.848699967 +0000 UTC m=+1974.417079236" Dec 03 09:10:54 crc kubenswrapper[4573]: I1203 09:10:54.058748 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e8386e5-315f-4d6f-a560-e1a45777e339" path="/var/lib/kubelet/pods/8e8386e5-315f-4d6f-a560-e1a45777e339/volumes" Dec 03 09:10:59 crc kubenswrapper[4573]: I1203 09:10:59.030344 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:10:59 crc kubenswrapper[4573]: I1203 09:10:59.880801 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"300e60f96de70ba98f90e77595dce33621c17e38eac9f7c366d93366d0b0c462"} Dec 03 09:11:06 crc kubenswrapper[4573]: I1203 09:11:06.486746 4573 scope.go:117] "RemoveContainer" containerID="84dcfceb254e96342219e95412e54d29973b82e2655f344fb62ea90137bce299" Dec 03 09:11:06 crc kubenswrapper[4573]: I1203 09:11:06.547434 4573 scope.go:117] "RemoveContainer" containerID="efd263f20f857341d85f4238f1a9435b948d198a4e7f1b02d40e17485eb33a8b" Dec 03 09:11:12 crc kubenswrapper[4573]: I1203 09:11:12.047996 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-kvbg6"] Dec 03 09:11:12 crc kubenswrapper[4573]: I1203 09:11:12.062157 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-kvbg6"] Dec 03 09:11:14 crc kubenswrapper[4573]: I1203 09:11:14.045503 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df6a51aa-75ab-4bdc-9e22-6ba714772ecf" path="/var/lib/kubelet/pods/df6a51aa-75ab-4bdc-9e22-6ba714772ecf/volumes" Dec 03 09:11:17 crc kubenswrapper[4573]: I1203 09:11:17.063581 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-m4vz4"] Dec 03 09:11:17 crc kubenswrapper[4573]: I1203 09:11:17.081804 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-m4vz4"] Dec 03 09:11:18 crc kubenswrapper[4573]: I1203 09:11:18.045405 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e28fb330-9e1a-4480-9a2d-b059d8e1576d" path="/var/lib/kubelet/pods/e28fb330-9e1a-4480-9a2d-b059d8e1576d/volumes" Dec 03 09:11:35 crc kubenswrapper[4573]: I1203 09:11:35.036040 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-fnppb"] Dec 03 09:11:35 crc kubenswrapper[4573]: I1203 09:11:35.046641 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-fnppb"] Dec 03 09:11:36 crc kubenswrapper[4573]: I1203 09:11:36.046865 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd84894-b4fa-481f-8856-678218f7bad7" path="/var/lib/kubelet/pods/3cd84894-b4fa-481f-8856-678218f7bad7/volumes" Dec 03 09:11:38 crc kubenswrapper[4573]: I1203 09:11:38.090290 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-chs2b"] Dec 03 09:11:38 crc kubenswrapper[4573]: I1203 09:11:38.107973 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-chs2b"] Dec 03 09:11:40 crc kubenswrapper[4573]: I1203 09:11:40.048623 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c73103b-e4fd-4a54-ad2c-021db8042b13" path="/var/lib/kubelet/pods/5c73103b-e4fd-4a54-ad2c-021db8042b13/volumes" Dec 03 09:12:06 crc kubenswrapper[4573]: I1203 09:12:06.676347 4573 scope.go:117] "RemoveContainer" containerID="d3a2b62460461d5aace5db31648f52a51f42f6610b73ce62ce94612ef769eab8" Dec 03 09:12:06 crc kubenswrapper[4573]: I1203 09:12:06.741328 4573 scope.go:117] "RemoveContainer" containerID="a831b5f7a5bb305ff88ecadd859e4cf44257739ca607a0df313d47dc032d6cc2" Dec 03 09:12:06 crc kubenswrapper[4573]: I1203 09:12:06.797036 4573 scope.go:117] "RemoveContainer" containerID="eb1fea2c39e1971637b2fb8d393068ecb9d4e67e5655a0afecace4db4cd54ca6" Dec 03 09:12:06 crc kubenswrapper[4573]: I1203 09:12:06.848460 4573 scope.go:117] "RemoveContainer" containerID="52e20e2bad1b07d9d79d83194aac0f7db5989ed58ba13e71f2fc4e566ec38b26" Dec 03 09:12:57 crc kubenswrapper[4573]: I1203 09:12:57.061542 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-9rq88"] Dec 03 09:12:57 crc kubenswrapper[4573]: I1203 09:12:57.072510 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-2kcns"] Dec 03 09:12:57 crc kubenswrapper[4573]: I1203 09:12:57.083805 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-kp7fb"] Dec 03 09:12:57 crc kubenswrapper[4573]: I1203 09:12:57.092274 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-9rq88"] Dec 03 09:12:57 crc kubenswrapper[4573]: I1203 09:12:57.103334 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-kp7fb"] Dec 03 09:12:57 crc kubenswrapper[4573]: I1203 09:12:57.110555 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-2kcns"] Dec 03 09:12:57 crc kubenswrapper[4573]: I1203 09:12:57.236609 4573 generic.go:334] "Generic (PLEG): container finished" podID="960face3-051e-4d24-b7ae-8e8b0f13d14e" containerID="297f3294e1cc02316d1b7b52b73510727c2e2ad25d5ecda1d65caddeb2a63338" exitCode=0 Dec 03 09:12:57 crc kubenswrapper[4573]: I1203 09:12:57.236662 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" event={"ID":"960face3-051e-4d24-b7ae-8e8b0f13d14e","Type":"ContainerDied","Data":"297f3294e1cc02316d1b7b52b73510727c2e2ad25d5ecda1d65caddeb2a63338"} Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.050630 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="539b8651-7e63-416c-8428-83cd2e4b3813" path="/var/lib/kubelet/pods/539b8651-7e63-416c-8428-83cd2e4b3813/volumes" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.053660 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad404c4-6b5d-4d8b-8298-efce4c5d54bb" path="/var/lib/kubelet/pods/7ad404c4-6b5d-4d8b-8298-efce4c5d54bb/volumes" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.054933 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff27d27-5fa2-401b-95a2-c633aa025759" path="/var/lib/kubelet/pods/7ff27d27-5fa2-401b-95a2-c633aa025759/volumes" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.056189 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-b5d2-account-create-update-wr7tj"] Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.064519 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-00ab-account-create-update-9bqrf"] Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.075307 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-96ff-account-create-update-mt94j"] Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.085699 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-00ab-account-create-update-9bqrf"] Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.093478 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-96ff-account-create-update-mt94j"] Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.102821 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-b5d2-account-create-update-wr7tj"] Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.681998 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.796437 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-ssh-key\") pod \"960face3-051e-4d24-b7ae-8e8b0f13d14e\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.796656 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld725\" (UniqueName: \"kubernetes.io/projected/960face3-051e-4d24-b7ae-8e8b0f13d14e-kube-api-access-ld725\") pod \"960face3-051e-4d24-b7ae-8e8b0f13d14e\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.796692 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-inventory\") pod \"960face3-051e-4d24-b7ae-8e8b0f13d14e\" (UID: \"960face3-051e-4d24-b7ae-8e8b0f13d14e\") " Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.810543 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/960face3-051e-4d24-b7ae-8e8b0f13d14e-kube-api-access-ld725" (OuterVolumeSpecName: "kube-api-access-ld725") pod "960face3-051e-4d24-b7ae-8e8b0f13d14e" (UID: "960face3-051e-4d24-b7ae-8e8b0f13d14e"). InnerVolumeSpecName "kube-api-access-ld725". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.825644 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-inventory" (OuterVolumeSpecName: "inventory") pod "960face3-051e-4d24-b7ae-8e8b0f13d14e" (UID: "960face3-051e-4d24-b7ae-8e8b0f13d14e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.834561 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "960face3-051e-4d24-b7ae-8e8b0f13d14e" (UID: "960face3-051e-4d24-b7ae-8e8b0f13d14e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.899810 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld725\" (UniqueName: \"kubernetes.io/projected/960face3-051e-4d24-b7ae-8e8b0f13d14e-kube-api-access-ld725\") on node \"crc\" DevicePath \"\"" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.900143 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:12:58 crc kubenswrapper[4573]: I1203 09:12:58.900291 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/960face3-051e-4d24-b7ae-8e8b0f13d14e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.256276 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" event={"ID":"960face3-051e-4d24-b7ae-8e8b0f13d14e","Type":"ContainerDied","Data":"02665d640b27316fd195ded4c56a9cc7e5bdfd19ef4ea1c29bfeda0696c065fe"} Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.256631 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02665d640b27316fd195ded4c56a9cc7e5bdfd19ef4ea1c29bfeda0696c065fe" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.256406 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6kp67" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.401881 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8"] Dec 03 09:12:59 crc kubenswrapper[4573]: E1203 09:12:59.403350 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="960face3-051e-4d24-b7ae-8e8b0f13d14e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.403381 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="960face3-051e-4d24-b7ae-8e8b0f13d14e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.411358 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="960face3-051e-4d24-b7ae-8e8b0f13d14e" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.412394 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.415010 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8"] Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.422446 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.422763 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.425121 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.425302 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.512152 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.512289 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kslqs\" (UniqueName: \"kubernetes.io/projected/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-kube-api-access-kslqs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.512332 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.614718 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kslqs\" (UniqueName: \"kubernetes.io/projected/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-kube-api-access-kslqs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.614812 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.614964 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.631094 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.631138 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.636998 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kslqs\" (UniqueName: \"kubernetes.io/projected/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-kube-api-access-kslqs\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:12:59 crc kubenswrapper[4573]: I1203 09:12:59.739706 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:13:00 crc kubenswrapper[4573]: I1203 09:13:00.059979 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f85ea37-8ca3-4b36-9c98-d1b2a51d108e" path="/var/lib/kubelet/pods/0f85ea37-8ca3-4b36-9c98-d1b2a51d108e/volumes" Dec 03 09:13:00 crc kubenswrapper[4573]: I1203 09:13:00.061191 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d287bf0-5630-48c1-89c7-2c3682991e6d" path="/var/lib/kubelet/pods/3d287bf0-5630-48c1-89c7-2c3682991e6d/volumes" Dec 03 09:13:00 crc kubenswrapper[4573]: I1203 09:13:00.061977 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0" path="/var/lib/kubelet/pods/5e2ebf6a-1287-48f4-a13d-4a31ae5c10e0/volumes" Dec 03 09:13:00 crc kubenswrapper[4573]: I1203 09:13:00.284986 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8"] Dec 03 09:13:00 crc kubenswrapper[4573]: I1203 09:13:00.485427 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:13:01 crc kubenswrapper[4573]: I1203 09:13:01.280638 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" event={"ID":"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5","Type":"ContainerStarted","Data":"2849fd51b97c7b4af1824053beda6d2bfa909bdc838305e02171609bf2ffd32d"} Dec 03 09:13:01 crc kubenswrapper[4573]: I1203 09:13:01.281018 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" event={"ID":"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5","Type":"ContainerStarted","Data":"f023013c53fce4f8750a1297162ad4441eebf91282c0d3411ffdd5b0a14ab20f"} Dec 03 09:13:01 crc kubenswrapper[4573]: I1203 09:13:01.307100 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" podStartSLOduration=2.119295397 podStartE2EDuration="2.307079094s" podCreationTimestamp="2025-12-03 09:12:59 +0000 UTC" firstStartedPulling="2025-12-03 09:13:00.294325805 +0000 UTC m=+2100.862705064" lastFinishedPulling="2025-12-03 09:13:00.482109492 +0000 UTC m=+2101.050488761" observedRunningTime="2025-12-03 09:13:01.30433773 +0000 UTC m=+2101.872716999" watchObservedRunningTime="2025-12-03 09:13:01.307079094 +0000 UTC m=+2101.875458353" Dec 03 09:13:06 crc kubenswrapper[4573]: I1203 09:13:06.971906 4573 scope.go:117] "RemoveContainer" containerID="720b8a43dea71e416a2c2843eaa0945696e842d829ec3f58a6988e63bf047368" Dec 03 09:13:07 crc kubenswrapper[4573]: I1203 09:13:07.006297 4573 scope.go:117] "RemoveContainer" containerID="340b52b445e91622c47165bd2cbbdd1cffe6096bbce9a1af515a640646a9312a" Dec 03 09:13:07 crc kubenswrapper[4573]: I1203 09:13:07.084099 4573 scope.go:117] "RemoveContainer" containerID="4988230dc7c3ce4c513117a7eb23941921f99b876e24ee199038a3618c4f4373" Dec 03 09:13:07 crc kubenswrapper[4573]: I1203 09:13:07.183988 4573 scope.go:117] "RemoveContainer" containerID="6b56c6795993e4802102a503d6b4bdb93cd5a0adb04f492a4a89db636958a5f9" Dec 03 09:13:07 crc kubenswrapper[4573]: I1203 09:13:07.217708 4573 scope.go:117] "RemoveContainer" containerID="fb0eb86bfd54149afc35d20883842b27700bb65d2b26f1a86893e9f54c787f13" Dec 03 09:13:07 crc kubenswrapper[4573]: I1203 09:13:07.270562 4573 scope.go:117] "RemoveContainer" containerID="a0f77cc7444581881aaea554f7a0c8d6b00b893c572ea1029e9064079cae3832" Dec 03 09:13:26 crc kubenswrapper[4573]: I1203 09:13:26.943540 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:13:26 crc kubenswrapper[4573]: I1203 09:13:26.944137 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:13:31 crc kubenswrapper[4573]: I1203 09:13:31.064516 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-22zjj"] Dec 03 09:13:31 crc kubenswrapper[4573]: I1203 09:13:31.081887 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-22zjj"] Dec 03 09:13:32 crc kubenswrapper[4573]: I1203 09:13:32.080383 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc" path="/var/lib/kubelet/pods/4b4e8f12-4f4b-4ae7-a18b-2fe2dd9427cc/volumes" Dec 03 09:13:56 crc kubenswrapper[4573]: I1203 09:13:56.943511 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:13:56 crc kubenswrapper[4573]: I1203 09:13:56.944248 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:14:04 crc kubenswrapper[4573]: I1203 09:14:04.057346 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s9sms"] Dec 03 09:14:04 crc kubenswrapper[4573]: I1203 09:14:04.070152 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-vmz2j"] Dec 03 09:14:04 crc kubenswrapper[4573]: I1203 09:14:04.085340 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s9sms"] Dec 03 09:14:04 crc kubenswrapper[4573]: I1203 09:14:04.094572 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-vmz2j"] Dec 03 09:14:06 crc kubenswrapper[4573]: I1203 09:14:06.044934 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3e275e6-2937-4c08-8b15-537b25e00cb1" path="/var/lib/kubelet/pods/f3e275e6-2937-4c08-8b15-537b25e00cb1/volumes" Dec 03 09:14:06 crc kubenswrapper[4573]: I1203 09:14:06.046493 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44" path="/var/lib/kubelet/pods/f4b6fbd4-2aa4-4f39-aa46-6e7a979aed44/volumes" Dec 03 09:14:07 crc kubenswrapper[4573]: I1203 09:14:07.474474 4573 scope.go:117] "RemoveContainer" containerID="e5abf83ae2290120529b2027ef11708338276a01b6082bda89e2b334c973e1ac" Dec 03 09:14:07 crc kubenswrapper[4573]: I1203 09:14:07.553146 4573 scope.go:117] "RemoveContainer" containerID="5b31ec267d0fa581174e02bec19006b16d2a3f12392424e35d0bad822f3acf27" Dec 03 09:14:07 crc kubenswrapper[4573]: I1203 09:14:07.601521 4573 scope.go:117] "RemoveContainer" containerID="f050e023d9d1085a72c09420178f8232c10a3acb16222c9a30a70643ea37e1c4" Dec 03 09:14:17 crc kubenswrapper[4573]: I1203 09:14:17.931082 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t6r2r"] Dec 03 09:14:17 crc kubenswrapper[4573]: I1203 09:14:17.934031 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:17 crc kubenswrapper[4573]: I1203 09:14:17.968322 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t6r2r"] Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.085983 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-utilities\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.086523 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6c8\" (UniqueName: \"kubernetes.io/projected/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-kube-api-access-7c6c8\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.086613 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-catalog-content\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.188602 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-utilities\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.188721 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6c8\" (UniqueName: \"kubernetes.io/projected/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-kube-api-access-7c6c8\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.188794 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-catalog-content\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.189484 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-catalog-content\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.189797 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-utilities\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.223929 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6c8\" (UniqueName: \"kubernetes.io/projected/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-kube-api-access-7c6c8\") pod \"certified-operators-t6r2r\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.273945 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:18 crc kubenswrapper[4573]: I1203 09:14:18.937023 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t6r2r"] Dec 03 09:14:19 crc kubenswrapper[4573]: I1203 09:14:19.607209 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6r2r" event={"ID":"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5","Type":"ContainerDied","Data":"e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278"} Dec 03 09:14:19 crc kubenswrapper[4573]: I1203 09:14:19.607039 4573 generic.go:334] "Generic (PLEG): container finished" podID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerID="e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278" exitCode=0 Dec 03 09:14:19 crc kubenswrapper[4573]: I1203 09:14:19.608026 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6r2r" event={"ID":"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5","Type":"ContainerStarted","Data":"402921223c594b526ef60e95b4d170429d5597aca38a184c818e339aa20e6933"} Dec 03 09:14:20 crc kubenswrapper[4573]: I1203 09:14:20.620884 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6r2r" event={"ID":"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5","Type":"ContainerStarted","Data":"d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921"} Dec 03 09:14:21 crc kubenswrapper[4573]: E1203 09:14:21.359619 4573 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94e2105f_4216_4ddc_b72e_6c69c1bdd0e5.slice/crio-d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921.scope\": RecentStats: unable to find data in memory cache]" Dec 03 09:14:21 crc kubenswrapper[4573]: I1203 09:14:21.635559 4573 generic.go:334] "Generic (PLEG): container finished" podID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerID="d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921" exitCode=0 Dec 03 09:14:21 crc kubenswrapper[4573]: I1203 09:14:21.635649 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6r2r" event={"ID":"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5","Type":"ContainerDied","Data":"d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921"} Dec 03 09:14:22 crc kubenswrapper[4573]: I1203 09:14:22.646927 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6r2r" event={"ID":"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5","Type":"ContainerStarted","Data":"0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45"} Dec 03 09:14:24 crc kubenswrapper[4573]: I1203 09:14:24.674992 4573 generic.go:334] "Generic (PLEG): container finished" podID="ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5" containerID="2849fd51b97c7b4af1824053beda6d2bfa909bdc838305e02171609bf2ffd32d" exitCode=0 Dec 03 09:14:24 crc kubenswrapper[4573]: I1203 09:14:24.675233 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" event={"ID":"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5","Type":"ContainerDied","Data":"2849fd51b97c7b4af1824053beda6d2bfa909bdc838305e02171609bf2ffd32d"} Dec 03 09:14:24 crc kubenswrapper[4573]: I1203 09:14:24.703891 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t6r2r" podStartSLOduration=5.000259106 podStartE2EDuration="7.703865045s" podCreationTimestamp="2025-12-03 09:14:17 +0000 UTC" firstStartedPulling="2025-12-03 09:14:19.609799095 +0000 UTC m=+2180.178178364" lastFinishedPulling="2025-12-03 09:14:22.313405044 +0000 UTC m=+2182.881784303" observedRunningTime="2025-12-03 09:14:22.683257301 +0000 UTC m=+2183.251636560" watchObservedRunningTime="2025-12-03 09:14:24.703865045 +0000 UTC m=+2185.272244304" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.231935 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.349332 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-ssh-key\") pod \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.349487 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kslqs\" (UniqueName: \"kubernetes.io/projected/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-kube-api-access-kslqs\") pod \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.349581 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-inventory\") pod \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\" (UID: \"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5\") " Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.358499 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-kube-api-access-kslqs" (OuterVolumeSpecName: "kube-api-access-kslqs") pod "ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5" (UID: "ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5"). InnerVolumeSpecName "kube-api-access-kslqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.385558 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5" (UID: "ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.387276 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-inventory" (OuterVolumeSpecName: "inventory") pod "ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5" (UID: "ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.452152 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.452189 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.452204 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kslqs\" (UniqueName: \"kubernetes.io/projected/ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5-kube-api-access-kslqs\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.698352 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" event={"ID":"ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5","Type":"ContainerDied","Data":"f023013c53fce4f8750a1297162ad4441eebf91282c0d3411ffdd5b0a14ab20f"} Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.698400 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f023013c53fce4f8750a1297162ad4441eebf91282c0d3411ffdd5b0a14ab20f" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.698475 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.827687 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf"] Dec 03 09:14:26 crc kubenswrapper[4573]: E1203 09:14:26.828284 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.828310 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.828604 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.829691 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.833543 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.834232 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.834762 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.837966 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf"] Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.850540 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.943947 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.944381 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.944506 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.945687 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"300e60f96de70ba98f90e77595dce33621c17e38eac9f7c366d93366d0b0c462"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.945856 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://300e60f96de70ba98f90e77595dce33621c17e38eac9f7c366d93366d0b0c462" gracePeriod=600 Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.969508 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.969560 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:26 crc kubenswrapper[4573]: I1203 09:14:26.969651 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92trn\" (UniqueName: \"kubernetes.io/projected/8ebf9479-f67b-443c-8ae2-3bec8e719750-kube-api-access-92trn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.071878 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.071933 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.071968 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92trn\" (UniqueName: \"kubernetes.io/projected/8ebf9479-f67b-443c-8ae2-3bec8e719750-kube-api-access-92trn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.085291 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.085848 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.093517 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92trn\" (UniqueName: \"kubernetes.io/projected/8ebf9479-f67b-443c-8ae2-3bec8e719750-kube-api-access-92trn\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.152200 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.717801 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="300e60f96de70ba98f90e77595dce33621c17e38eac9f7c366d93366d0b0c462" exitCode=0 Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.718264 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"300e60f96de70ba98f90e77595dce33621c17e38eac9f7c366d93366d0b0c462"} Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.718356 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3"} Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.718395 4573 scope.go:117] "RemoveContainer" containerID="4e6f7a3bc829856f6aca0cee2b36f9247b58aefc8e4116f7b4ed08432741ce85" Dec 03 09:14:27 crc kubenswrapper[4573]: I1203 09:14:27.873262 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf"] Dec 03 09:14:28 crc kubenswrapper[4573]: I1203 09:14:28.275191 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:28 crc kubenswrapper[4573]: I1203 09:14:28.275739 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:28 crc kubenswrapper[4573]: I1203 09:14:28.344833 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:28 crc kubenswrapper[4573]: I1203 09:14:28.732974 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" event={"ID":"8ebf9479-f67b-443c-8ae2-3bec8e719750","Type":"ContainerStarted","Data":"b0e9fbde03e2de8e259bdbc003264115720c90eae7867420b5fb65afa3365b77"} Dec 03 09:14:28 crc kubenswrapper[4573]: I1203 09:14:28.734767 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" event={"ID":"8ebf9479-f67b-443c-8ae2-3bec8e719750","Type":"ContainerStarted","Data":"8e800abb99a0c99281be2155d604cd08d6ce626f2abeece226fc008210c6d1f9"} Dec 03 09:14:28 crc kubenswrapper[4573]: I1203 09:14:28.762838 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" podStartSLOduration=2.581310222 podStartE2EDuration="2.762807977s" podCreationTimestamp="2025-12-03 09:14:26 +0000 UTC" firstStartedPulling="2025-12-03 09:14:27.884657133 +0000 UTC m=+2188.453036392" lastFinishedPulling="2025-12-03 09:14:28.066154888 +0000 UTC m=+2188.634534147" observedRunningTime="2025-12-03 09:14:28.746816951 +0000 UTC m=+2189.315196210" watchObservedRunningTime="2025-12-03 09:14:28.762807977 +0000 UTC m=+2189.331187236" Dec 03 09:14:28 crc kubenswrapper[4573]: I1203 09:14:28.786562 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:28 crc kubenswrapper[4573]: I1203 09:14:28.846438 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t6r2r"] Dec 03 09:14:30 crc kubenswrapper[4573]: I1203 09:14:30.753230 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t6r2r" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerName="registry-server" containerID="cri-o://0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45" gracePeriod=2 Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.321857 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.429187 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-catalog-content\") pod \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.429270 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c6c8\" (UniqueName: \"kubernetes.io/projected/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-kube-api-access-7c6c8\") pod \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.429322 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-utilities\") pod \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\" (UID: \"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5\") " Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.430480 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-utilities" (OuterVolumeSpecName: "utilities") pod "94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" (UID: "94e2105f-4216-4ddc-b72e-6c69c1bdd0e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.439417 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-kube-api-access-7c6c8" (OuterVolumeSpecName: "kube-api-access-7c6c8") pod "94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" (UID: "94e2105f-4216-4ddc-b72e-6c69c1bdd0e5"). InnerVolumeSpecName "kube-api-access-7c6c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.482846 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" (UID: "94e2105f-4216-4ddc-b72e-6c69c1bdd0e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.532254 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.532292 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c6c8\" (UniqueName: \"kubernetes.io/projected/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-kube-api-access-7c6c8\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.532305 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.766594 4573 generic.go:334] "Generic (PLEG): container finished" podID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerID="0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45" exitCode=0 Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.766669 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t6r2r" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.766715 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6r2r" event={"ID":"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5","Type":"ContainerDied","Data":"0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45"} Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.768152 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t6r2r" event={"ID":"94e2105f-4216-4ddc-b72e-6c69c1bdd0e5","Type":"ContainerDied","Data":"402921223c594b526ef60e95b4d170429d5597aca38a184c818e339aa20e6933"} Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.768233 4573 scope.go:117] "RemoveContainer" containerID="0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.803645 4573 scope.go:117] "RemoveContainer" containerID="d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.833762 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t6r2r"] Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.844573 4573 scope.go:117] "RemoveContainer" containerID="e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.847430 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t6r2r"] Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.910632 4573 scope.go:117] "RemoveContainer" containerID="0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45" Dec 03 09:14:31 crc kubenswrapper[4573]: E1203 09:14:31.911217 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45\": container with ID starting with 0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45 not found: ID does not exist" containerID="0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.911253 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45"} err="failed to get container status \"0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45\": rpc error: code = NotFound desc = could not find container \"0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45\": container with ID starting with 0949907050d73fd4a43bba264141c35ec5182ce83dacbcd81d1ddaa8b86e3a45 not found: ID does not exist" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.911301 4573 scope.go:117] "RemoveContainer" containerID="d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921" Dec 03 09:14:31 crc kubenswrapper[4573]: E1203 09:14:31.911646 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921\": container with ID starting with d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921 not found: ID does not exist" containerID="d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.911770 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921"} err="failed to get container status \"d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921\": rpc error: code = NotFound desc = could not find container \"d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921\": container with ID starting with d9bbd8d7646cb678e19589485b16b7e71d39eb5c19bc7bef05a7b02f7c606921 not found: ID does not exist" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.911867 4573 scope.go:117] "RemoveContainer" containerID="e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278" Dec 03 09:14:31 crc kubenswrapper[4573]: E1203 09:14:31.912265 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278\": container with ID starting with e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278 not found: ID does not exist" containerID="e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278" Dec 03 09:14:31 crc kubenswrapper[4573]: I1203 09:14:31.912292 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278"} err="failed to get container status \"e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278\": rpc error: code = NotFound desc = could not find container \"e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278\": container with ID starting with e74c44e001642d381734660c5cba93a140b1f902748114bbd678fb7bbf2fd278 not found: ID does not exist" Dec 03 09:14:32 crc kubenswrapper[4573]: I1203 09:14:32.044783 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" path="/var/lib/kubelet/pods/94e2105f-4216-4ddc-b72e-6c69c1bdd0e5/volumes" Dec 03 09:14:33 crc kubenswrapper[4573]: I1203 09:14:33.786748 4573 generic.go:334] "Generic (PLEG): container finished" podID="8ebf9479-f67b-443c-8ae2-3bec8e719750" containerID="b0e9fbde03e2de8e259bdbc003264115720c90eae7867420b5fb65afa3365b77" exitCode=0 Dec 03 09:14:33 crc kubenswrapper[4573]: I1203 09:14:33.786825 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" event={"ID":"8ebf9479-f67b-443c-8ae2-3bec8e719750","Type":"ContainerDied","Data":"b0e9fbde03e2de8e259bdbc003264115720c90eae7867420b5fb65afa3365b77"} Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.015114 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h74lz"] Dec 03 09:14:34 crc kubenswrapper[4573]: E1203 09:14:34.015581 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerName="extract-utilities" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.015602 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerName="extract-utilities" Dec 03 09:14:34 crc kubenswrapper[4573]: E1203 09:14:34.015650 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerName="registry-server" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.015657 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerName="registry-server" Dec 03 09:14:34 crc kubenswrapper[4573]: E1203 09:14:34.015673 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerName="extract-content" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.015678 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerName="extract-content" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.015887 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e2105f-4216-4ddc-b72e-6c69c1bdd0e5" containerName="registry-server" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.017303 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.027910 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h74lz"] Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.186247 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-catalog-content\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.186368 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-utilities\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.186476 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq927\" (UniqueName: \"kubernetes.io/projected/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-kube-api-access-zq927\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.288504 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-utilities\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.288603 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq927\" (UniqueName: \"kubernetes.io/projected/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-kube-api-access-zq927\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.288681 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-catalog-content\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.289201 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-catalog-content\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.289260 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-utilities\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.316656 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq927\" (UniqueName: \"kubernetes.io/projected/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-kube-api-access-zq927\") pod \"redhat-operators-h74lz\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.359883 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:34 crc kubenswrapper[4573]: I1203 09:14:34.943035 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h74lz"] Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.445679 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.535739 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-ssh-key\") pod \"8ebf9479-f67b-443c-8ae2-3bec8e719750\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.536086 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-inventory\") pod \"8ebf9479-f67b-443c-8ae2-3bec8e719750\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.536696 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92trn\" (UniqueName: \"kubernetes.io/projected/8ebf9479-f67b-443c-8ae2-3bec8e719750-kube-api-access-92trn\") pod \"8ebf9479-f67b-443c-8ae2-3bec8e719750\" (UID: \"8ebf9479-f67b-443c-8ae2-3bec8e719750\") " Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.570960 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ebf9479-f67b-443c-8ae2-3bec8e719750-kube-api-access-92trn" (OuterVolumeSpecName: "kube-api-access-92trn") pod "8ebf9479-f67b-443c-8ae2-3bec8e719750" (UID: "8ebf9479-f67b-443c-8ae2-3bec8e719750"). InnerVolumeSpecName "kube-api-access-92trn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.576331 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-inventory" (OuterVolumeSpecName: "inventory") pod "8ebf9479-f67b-443c-8ae2-3bec8e719750" (UID: "8ebf9479-f67b-443c-8ae2-3bec8e719750"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.578273 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8ebf9479-f67b-443c-8ae2-3bec8e719750" (UID: "8ebf9479-f67b-443c-8ae2-3bec8e719750"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.637696 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92trn\" (UniqueName: \"kubernetes.io/projected/8ebf9479-f67b-443c-8ae2-3bec8e719750-kube-api-access-92trn\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.637732 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.637743 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ebf9479-f67b-443c-8ae2-3bec8e719750-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.805687 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" event={"ID":"8ebf9479-f67b-443c-8ae2-3bec8e719750","Type":"ContainerDied","Data":"8e800abb99a0c99281be2155d604cd08d6ce626f2abeece226fc008210c6d1f9"} Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.805750 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e800abb99a0c99281be2155d604cd08d6ce626f2abeece226fc008210c6d1f9" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.806083 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.808018 4573 generic.go:334] "Generic (PLEG): container finished" podID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerID="e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4" exitCode=0 Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.808194 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h74lz" event={"ID":"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1","Type":"ContainerDied","Data":"e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4"} Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.808238 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h74lz" event={"ID":"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1","Type":"ContainerStarted","Data":"4544eb44613c498569b7856942d3ad807c4802e80b8ac735da777843509b057b"} Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.927136 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8"] Dec 03 09:14:35 crc kubenswrapper[4573]: E1203 09:14:35.927626 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ebf9479-f67b-443c-8ae2-3bec8e719750" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.927652 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ebf9479-f67b-443c-8ae2-3bec8e719750" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.927913 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ebf9479-f67b-443c-8ae2-3bec8e719750" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.928768 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.931218 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.932659 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.932926 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.933469 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.943974 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.944136 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tndcw\" (UniqueName: \"kubernetes.io/projected/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-kube-api-access-tndcw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.944307 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:35 crc kubenswrapper[4573]: I1203 09:14:35.955225 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8"] Dec 03 09:14:36 crc kubenswrapper[4573]: I1203 09:14:36.046555 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:36 crc kubenswrapper[4573]: I1203 09:14:36.046901 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:36 crc kubenswrapper[4573]: I1203 09:14:36.047014 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tndcw\" (UniqueName: \"kubernetes.io/projected/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-kube-api-access-tndcw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:36 crc kubenswrapper[4573]: I1203 09:14:36.051554 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:36 crc kubenswrapper[4573]: I1203 09:14:36.063946 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:36 crc kubenswrapper[4573]: I1203 09:14:36.068727 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tndcw\" (UniqueName: \"kubernetes.io/projected/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-kube-api-access-tndcw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cg5z8\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:36 crc kubenswrapper[4573]: I1203 09:14:36.246283 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:14:36 crc kubenswrapper[4573]: I1203 09:14:36.889476 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8"] Dec 03 09:14:37 crc kubenswrapper[4573]: I1203 09:14:37.838600 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h74lz" event={"ID":"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1","Type":"ContainerStarted","Data":"1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3"} Dec 03 09:14:37 crc kubenswrapper[4573]: I1203 09:14:37.847455 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" event={"ID":"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b","Type":"ContainerStarted","Data":"e04738bc3c16ad522b1092644cef2da4acdb3566322aeeb5c595ab80ce56968c"} Dec 03 09:14:37 crc kubenswrapper[4573]: I1203 09:14:37.847490 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" event={"ID":"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b","Type":"ContainerStarted","Data":"ef98f9ea84096baa000d304edd1d950f779ca66fff5ca438b089e6bf203f14b8"} Dec 03 09:14:37 crc kubenswrapper[4573]: I1203 09:14:37.905104 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" podStartSLOduration=2.742765506 podStartE2EDuration="2.905078207s" podCreationTimestamp="2025-12-03 09:14:35 +0000 UTC" firstStartedPulling="2025-12-03 09:14:36.899399691 +0000 UTC m=+2197.467778950" lastFinishedPulling="2025-12-03 09:14:37.061712392 +0000 UTC m=+2197.630091651" observedRunningTime="2025-12-03 09:14:37.890640892 +0000 UTC m=+2198.459020151" watchObservedRunningTime="2025-12-03 09:14:37.905078207 +0000 UTC m=+2198.473457466" Dec 03 09:14:40 crc kubenswrapper[4573]: I1203 09:14:40.886620 4573 generic.go:334] "Generic (PLEG): container finished" podID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerID="1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3" exitCode=0 Dec 03 09:14:40 crc kubenswrapper[4573]: I1203 09:14:40.886721 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h74lz" event={"ID":"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1","Type":"ContainerDied","Data":"1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3"} Dec 03 09:14:41 crc kubenswrapper[4573]: I1203 09:14:41.899275 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h74lz" event={"ID":"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1","Type":"ContainerStarted","Data":"1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552"} Dec 03 09:14:41 crc kubenswrapper[4573]: I1203 09:14:41.926323 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h74lz" podStartSLOduration=3.305475547 podStartE2EDuration="8.926293978s" podCreationTimestamp="2025-12-03 09:14:33 +0000 UTC" firstStartedPulling="2025-12-03 09:14:35.809897407 +0000 UTC m=+2196.378276666" lastFinishedPulling="2025-12-03 09:14:41.430715838 +0000 UTC m=+2201.999095097" observedRunningTime="2025-12-03 09:14:41.921186218 +0000 UTC m=+2202.489565507" watchObservedRunningTime="2025-12-03 09:14:41.926293978 +0000 UTC m=+2202.494673237" Dec 03 09:14:43 crc kubenswrapper[4573]: I1203 09:14:43.042659 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-pgtv2"] Dec 03 09:14:43 crc kubenswrapper[4573]: I1203 09:14:43.056092 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-pgtv2"] Dec 03 09:14:44 crc kubenswrapper[4573]: I1203 09:14:44.044139 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="178dc07a-005e-4e3d-9c12-28b42eb2cc7d" path="/var/lib/kubelet/pods/178dc07a-005e-4e3d-9c12-28b42eb2cc7d/volumes" Dec 03 09:14:44 crc kubenswrapper[4573]: I1203 09:14:44.360155 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:44 crc kubenswrapper[4573]: I1203 09:14:44.360211 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:45 crc kubenswrapper[4573]: I1203 09:14:45.416407 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h74lz" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="registry-server" probeResult="failure" output=< Dec 03 09:14:45 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:14:45 crc kubenswrapper[4573]: > Dec 03 09:14:54 crc kubenswrapper[4573]: I1203 09:14:54.411673 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:54 crc kubenswrapper[4573]: I1203 09:14:54.468764 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:54 crc kubenswrapper[4573]: I1203 09:14:54.668882 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h74lz"] Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.030774 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h74lz" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="registry-server" containerID="cri-o://1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552" gracePeriod=2 Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.584174 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.674443 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq927\" (UniqueName: \"kubernetes.io/projected/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-kube-api-access-zq927\") pod \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.674510 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-utilities\") pod \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.674653 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-catalog-content\") pod \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\" (UID: \"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1\") " Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.675588 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-utilities" (OuterVolumeSpecName: "utilities") pod "7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" (UID: "7fe83c77-8995-4db6-bfd0-d47afdc8b7e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.685374 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-kube-api-access-zq927" (OuterVolumeSpecName: "kube-api-access-zq927") pod "7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" (UID: "7fe83c77-8995-4db6-bfd0-d47afdc8b7e1"). InnerVolumeSpecName "kube-api-access-zq927". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.777508 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq927\" (UniqueName: \"kubernetes.io/projected/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-kube-api-access-zq927\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.777551 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.806143 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" (UID: "7fe83c77-8995-4db6-bfd0-d47afdc8b7e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:14:56 crc kubenswrapper[4573]: I1203 09:14:56.879887 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.043752 4573 generic.go:334] "Generic (PLEG): container finished" podID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerID="1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552" exitCode=0 Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.043801 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h74lz" event={"ID":"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1","Type":"ContainerDied","Data":"1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552"} Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.043833 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h74lz" event={"ID":"7fe83c77-8995-4db6-bfd0-d47afdc8b7e1","Type":"ContainerDied","Data":"4544eb44613c498569b7856942d3ad807c4802e80b8ac735da777843509b057b"} Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.043870 4573 scope.go:117] "RemoveContainer" containerID="1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.044022 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h74lz" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.092248 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h74lz"] Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.103360 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h74lz"] Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.118311 4573 scope.go:117] "RemoveContainer" containerID="1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.158937 4573 scope.go:117] "RemoveContainer" containerID="e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.215134 4573 scope.go:117] "RemoveContainer" containerID="1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552" Dec 03 09:14:57 crc kubenswrapper[4573]: E1203 09:14:57.215468 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552\": container with ID starting with 1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552 not found: ID does not exist" containerID="1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.216037 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552"} err="failed to get container status \"1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552\": rpc error: code = NotFound desc = could not find container \"1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552\": container with ID starting with 1e9eda8a01879521749dc78be93346b98f4f2039ac7c418c5ee6586d54d6c552 not found: ID does not exist" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.216087 4573 scope.go:117] "RemoveContainer" containerID="1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3" Dec 03 09:14:57 crc kubenswrapper[4573]: E1203 09:14:57.216379 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3\": container with ID starting with 1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3 not found: ID does not exist" containerID="1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.216403 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3"} err="failed to get container status \"1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3\": rpc error: code = NotFound desc = could not find container \"1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3\": container with ID starting with 1b40b1e5ee3534abc4704c3edb4d1f6612ecc84502c242c372e77e7052c959e3 not found: ID does not exist" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.216426 4573 scope.go:117] "RemoveContainer" containerID="e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4" Dec 03 09:14:57 crc kubenswrapper[4573]: E1203 09:14:57.216947 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4\": container with ID starting with e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4 not found: ID does not exist" containerID="e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4" Dec 03 09:14:57 crc kubenswrapper[4573]: I1203 09:14:57.216995 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4"} err="failed to get container status \"e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4\": rpc error: code = NotFound desc = could not find container \"e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4\": container with ID starting with e55d4ea8058e7c597333d6e1585d892cb7b7ae3c5eb7407c16cd7fccf2f9aae4 not found: ID does not exist" Dec 03 09:14:58 crc kubenswrapper[4573]: I1203 09:14:58.044289 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" path="/var/lib/kubelet/pods/7fe83c77-8995-4db6-bfd0-d47afdc8b7e1/volumes" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.157489 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn"] Dec 03 09:15:00 crc kubenswrapper[4573]: E1203 09:15:00.160256 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="extract-content" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.160279 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="extract-content" Dec 03 09:15:00 crc kubenswrapper[4573]: E1203 09:15:00.160313 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="registry-server" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.160320 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="registry-server" Dec 03 09:15:00 crc kubenswrapper[4573]: E1203 09:15:00.160339 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="extract-utilities" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.160345 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="extract-utilities" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.160534 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe83c77-8995-4db6-bfd0-d47afdc8b7e1" containerName="registry-server" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.161249 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.164870 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.165287 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.192247 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn"] Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.280866 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrv5k\" (UniqueName: \"kubernetes.io/projected/4dc0c400-a678-4df2-9e5b-f7163404f664-kube-api-access-nrv5k\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.280969 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc0c400-a678-4df2-9e5b-f7163404f664-secret-volume\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.281008 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc0c400-a678-4df2-9e5b-f7163404f664-config-volume\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.382969 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrv5k\" (UniqueName: \"kubernetes.io/projected/4dc0c400-a678-4df2-9e5b-f7163404f664-kube-api-access-nrv5k\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.383073 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc0c400-a678-4df2-9e5b-f7163404f664-secret-volume\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.383099 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc0c400-a678-4df2-9e5b-f7163404f664-config-volume\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.384064 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc0c400-a678-4df2-9e5b-f7163404f664-config-volume\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.402410 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrv5k\" (UniqueName: \"kubernetes.io/projected/4dc0c400-a678-4df2-9e5b-f7163404f664-kube-api-access-nrv5k\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.406833 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc0c400-a678-4df2-9e5b-f7163404f664-secret-volume\") pod \"collect-profiles-29412555-z8gxn\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:00 crc kubenswrapper[4573]: I1203 09:15:00.485760 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:01 crc kubenswrapper[4573]: I1203 09:15:01.013039 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn"] Dec 03 09:15:01 crc kubenswrapper[4573]: I1203 09:15:01.081459 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" event={"ID":"4dc0c400-a678-4df2-9e5b-f7163404f664","Type":"ContainerStarted","Data":"807421a64b008b07d01f2bdd6e8b6410b705f15b0dd5dde7befebaf643103538"} Dec 03 09:15:02 crc kubenswrapper[4573]: I1203 09:15:02.115166 4573 generic.go:334] "Generic (PLEG): container finished" podID="4dc0c400-a678-4df2-9e5b-f7163404f664" containerID="1fabd5357e3df26b564ad7aaea88f5939ff8ac24eec6d7e65dfcc406ddb2cf7b" exitCode=0 Dec 03 09:15:02 crc kubenswrapper[4573]: I1203 09:15:02.115465 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" event={"ID":"4dc0c400-a678-4df2-9e5b-f7163404f664","Type":"ContainerDied","Data":"1fabd5357e3df26b564ad7aaea88f5939ff8ac24eec6d7e65dfcc406ddb2cf7b"} Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.519276 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.679582 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc0c400-a678-4df2-9e5b-f7163404f664-config-volume\") pod \"4dc0c400-a678-4df2-9e5b-f7163404f664\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.679768 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrv5k\" (UniqueName: \"kubernetes.io/projected/4dc0c400-a678-4df2-9e5b-f7163404f664-kube-api-access-nrv5k\") pod \"4dc0c400-a678-4df2-9e5b-f7163404f664\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.680667 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dc0c400-a678-4df2-9e5b-f7163404f664-config-volume" (OuterVolumeSpecName: "config-volume") pod "4dc0c400-a678-4df2-9e5b-f7163404f664" (UID: "4dc0c400-a678-4df2-9e5b-f7163404f664"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.681363 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc0c400-a678-4df2-9e5b-f7163404f664-secret-volume\") pod \"4dc0c400-a678-4df2-9e5b-f7163404f664\" (UID: \"4dc0c400-a678-4df2-9e5b-f7163404f664\") " Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.682366 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc0c400-a678-4df2-9e5b-f7163404f664-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.687504 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dc0c400-a678-4df2-9e5b-f7163404f664-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4dc0c400-a678-4df2-9e5b-f7163404f664" (UID: "4dc0c400-a678-4df2-9e5b-f7163404f664"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.688175 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dc0c400-a678-4df2-9e5b-f7163404f664-kube-api-access-nrv5k" (OuterVolumeSpecName: "kube-api-access-nrv5k") pod "4dc0c400-a678-4df2-9e5b-f7163404f664" (UID: "4dc0c400-a678-4df2-9e5b-f7163404f664"). InnerVolumeSpecName "kube-api-access-nrv5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.784142 4573 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc0c400-a678-4df2-9e5b-f7163404f664-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:03 crc kubenswrapper[4573]: I1203 09:15:03.784190 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrv5k\" (UniqueName: \"kubernetes.io/projected/4dc0c400-a678-4df2-9e5b-f7163404f664-kube-api-access-nrv5k\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:04 crc kubenswrapper[4573]: I1203 09:15:04.132234 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" event={"ID":"4dc0c400-a678-4df2-9e5b-f7163404f664","Type":"ContainerDied","Data":"807421a64b008b07d01f2bdd6e8b6410b705f15b0dd5dde7befebaf643103538"} Dec 03 09:15:04 crc kubenswrapper[4573]: I1203 09:15:04.132284 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="807421a64b008b07d01f2bdd6e8b6410b705f15b0dd5dde7befebaf643103538" Dec 03 09:15:04 crc kubenswrapper[4573]: I1203 09:15:04.132330 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn" Dec 03 09:15:04 crc kubenswrapper[4573]: I1203 09:15:04.610422 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs"] Dec 03 09:15:04 crc kubenswrapper[4573]: I1203 09:15:04.619515 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412510-r8ngs"] Dec 03 09:15:06 crc kubenswrapper[4573]: I1203 09:15:06.051143 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c10cae9-fe93-4134-ae2f-7e242201dc5f" path="/var/lib/kubelet/pods/2c10cae9-fe93-4134-ae2f-7e242201dc5f/volumes" Dec 03 09:15:07 crc kubenswrapper[4573]: I1203 09:15:07.773823 4573 scope.go:117] "RemoveContainer" containerID="1bc7ef0526e408e034ba5fcf74b11310b9a35a4dc23966e1e088c96dd51881a8" Dec 03 09:15:07 crc kubenswrapper[4573]: I1203 09:15:07.813591 4573 scope.go:117] "RemoveContainer" containerID="04af37895750db5d16417f6b792e13afe02547877c30ce1dd17ba189134517bd" Dec 03 09:15:23 crc kubenswrapper[4573]: I1203 09:15:23.474323 4573 generic.go:334] "Generic (PLEG): container finished" podID="2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b" containerID="e04738bc3c16ad522b1092644cef2da4acdb3566322aeeb5c595ab80ce56968c" exitCode=0 Dec 03 09:15:23 crc kubenswrapper[4573]: I1203 09:15:23.474893 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" event={"ID":"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b","Type":"ContainerDied","Data":"e04738bc3c16ad522b1092644cef2da4acdb3566322aeeb5c595ab80ce56968c"} Dec 03 09:15:24 crc kubenswrapper[4573]: I1203 09:15:24.950999 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:15:24 crc kubenswrapper[4573]: I1203 09:15:24.971000 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tndcw\" (UniqueName: \"kubernetes.io/projected/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-kube-api-access-tndcw\") pod \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " Dec 03 09:15:24 crc kubenswrapper[4573]: I1203 09:15:24.971082 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-inventory\") pod \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " Dec 03 09:15:24 crc kubenswrapper[4573]: I1203 09:15:24.979889 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-kube-api-access-tndcw" (OuterVolumeSpecName: "kube-api-access-tndcw") pod "2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b" (UID: "2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b"). InnerVolumeSpecName "kube-api-access-tndcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.002838 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-inventory" (OuterVolumeSpecName: "inventory") pod "2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b" (UID: "2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.074209 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-ssh-key\") pod \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\" (UID: \"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b\") " Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.074801 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tndcw\" (UniqueName: \"kubernetes.io/projected/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-kube-api-access-tndcw\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.074832 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.110944 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b" (UID: "2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.177002 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.520692 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" event={"ID":"2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b","Type":"ContainerDied","Data":"ef98f9ea84096baa000d304edd1d950f779ca66fff5ca438b089e6bf203f14b8"} Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.521018 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef98f9ea84096baa000d304edd1d950f779ca66fff5ca438b089e6bf203f14b8" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.520743 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cg5z8" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.673122 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm"] Dec 03 09:15:25 crc kubenswrapper[4573]: E1203 09:15:25.673553 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.673570 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:15:25 crc kubenswrapper[4573]: E1203 09:15:25.673593 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc0c400-a678-4df2-9e5b-f7163404f664" containerName="collect-profiles" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.673600 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc0c400-a678-4df2-9e5b-f7163404f664" containerName="collect-profiles" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.673783 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.673809 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dc0c400-a678-4df2-9e5b-f7163404f664" containerName="collect-profiles" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.674498 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.682799 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm"] Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.684960 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.685342 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.685940 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.686170 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.789143 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.789243 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9jw8\" (UniqueName: \"kubernetes.io/projected/6ea550b6-a03c-42ff-b766-073b5ca7bf15-kube-api-access-n9jw8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.790116 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.891640 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.891718 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.891772 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9jw8\" (UniqueName: \"kubernetes.io/projected/6ea550b6-a03c-42ff-b766-073b5ca7bf15-kube-api-access-n9jw8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.895938 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.897324 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.911491 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9jw8\" (UniqueName: \"kubernetes.io/projected/6ea550b6-a03c-42ff-b766-073b5ca7bf15-kube-api-access-n9jw8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-swdkm\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:25 crc kubenswrapper[4573]: I1203 09:15:25.992893 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:15:26 crc kubenswrapper[4573]: I1203 09:15:26.604969 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm"] Dec 03 09:15:26 crc kubenswrapper[4573]: I1203 09:15:26.616601 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:15:27 crc kubenswrapper[4573]: I1203 09:15:27.549825 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" event={"ID":"6ea550b6-a03c-42ff-b766-073b5ca7bf15","Type":"ContainerStarted","Data":"03ffd0e0830dde9a25a2f0606a47b7e605e27de7561bb81b48413c468173247a"} Dec 03 09:15:27 crc kubenswrapper[4573]: I1203 09:15:27.550447 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" event={"ID":"6ea550b6-a03c-42ff-b766-073b5ca7bf15","Type":"ContainerStarted","Data":"c90112f532149c9496716342b0096185f86cdb34418b377004e2337bc38ad65e"} Dec 03 09:15:27 crc kubenswrapper[4573]: I1203 09:15:27.575123 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" podStartSLOduration=2.354139624 podStartE2EDuration="2.575101266s" podCreationTimestamp="2025-12-03 09:15:25 +0000 UTC" firstStartedPulling="2025-12-03 09:15:26.616160257 +0000 UTC m=+2247.184539536" lastFinishedPulling="2025-12-03 09:15:26.837121899 +0000 UTC m=+2247.405501178" observedRunningTime="2025-12-03 09:15:27.574812578 +0000 UTC m=+2248.143191867" watchObservedRunningTime="2025-12-03 09:15:27.575101266 +0000 UTC m=+2248.143480535" Dec 03 09:16:25 crc kubenswrapper[4573]: I1203 09:16:25.177442 4573 generic.go:334] "Generic (PLEG): container finished" podID="6ea550b6-a03c-42ff-b766-073b5ca7bf15" containerID="03ffd0e0830dde9a25a2f0606a47b7e605e27de7561bb81b48413c468173247a" exitCode=0 Dec 03 09:16:25 crc kubenswrapper[4573]: I1203 09:16:25.177656 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" event={"ID":"6ea550b6-a03c-42ff-b766-073b5ca7bf15","Type":"ContainerDied","Data":"03ffd0e0830dde9a25a2f0606a47b7e605e27de7561bb81b48413c468173247a"} Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.653526 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.762856 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9jw8\" (UniqueName: \"kubernetes.io/projected/6ea550b6-a03c-42ff-b766-073b5ca7bf15-kube-api-access-n9jw8\") pod \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.763317 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-inventory\") pod \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.763403 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-ssh-key\") pod \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\" (UID: \"6ea550b6-a03c-42ff-b766-073b5ca7bf15\") " Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.770278 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea550b6-a03c-42ff-b766-073b5ca7bf15-kube-api-access-n9jw8" (OuterVolumeSpecName: "kube-api-access-n9jw8") pod "6ea550b6-a03c-42ff-b766-073b5ca7bf15" (UID: "6ea550b6-a03c-42ff-b766-073b5ca7bf15"). InnerVolumeSpecName "kube-api-access-n9jw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.792402 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6ea550b6-a03c-42ff-b766-073b5ca7bf15" (UID: "6ea550b6-a03c-42ff-b766-073b5ca7bf15"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.816670 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-inventory" (OuterVolumeSpecName: "inventory") pod "6ea550b6-a03c-42ff-b766-073b5ca7bf15" (UID: "6ea550b6-a03c-42ff-b766-073b5ca7bf15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.867354 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9jw8\" (UniqueName: \"kubernetes.io/projected/6ea550b6-a03c-42ff-b766-073b5ca7bf15-kube-api-access-n9jw8\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.867412 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:26 crc kubenswrapper[4573]: I1203 09:16:26.867433 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6ea550b6-a03c-42ff-b766-073b5ca7bf15-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.205962 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" event={"ID":"6ea550b6-a03c-42ff-b766-073b5ca7bf15","Type":"ContainerDied","Data":"c90112f532149c9496716342b0096185f86cdb34418b377004e2337bc38ad65e"} Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.206018 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c90112f532149c9496716342b0096185f86cdb34418b377004e2337bc38ad65e" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.206033 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-swdkm" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.392246 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gc8nt"] Dec 03 09:16:27 crc kubenswrapper[4573]: E1203 09:16:27.392680 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea550b6-a03c-42ff-b766-073b5ca7bf15" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.392698 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea550b6-a03c-42ff-b766-073b5ca7bf15" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.392897 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea550b6-a03c-42ff-b766-073b5ca7bf15" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.393592 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.395971 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.396381 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.396652 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.400632 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.407282 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gc8nt"] Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.478275 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx5t8\" (UniqueName: \"kubernetes.io/projected/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-kube-api-access-xx5t8\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.478446 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.478577 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.579494 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx5t8\" (UniqueName: \"kubernetes.io/projected/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-kube-api-access-xx5t8\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.579782 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.579849 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.585786 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.586812 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.596740 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx5t8\" (UniqueName: \"kubernetes.io/projected/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-kube-api-access-xx5t8\") pod \"ssh-known-hosts-edpm-deployment-gc8nt\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:27 crc kubenswrapper[4573]: I1203 09:16:27.754835 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:28 crc kubenswrapper[4573]: I1203 09:16:28.359996 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-gc8nt"] Dec 03 09:16:29 crc kubenswrapper[4573]: I1203 09:16:29.226136 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" event={"ID":"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3","Type":"ContainerStarted","Data":"ad8608f6aa3b9ebf9287e1a4f7358de86e308537533378b1062a4fd1fb35d986"} Dec 03 09:16:29 crc kubenswrapper[4573]: I1203 09:16:29.226653 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" event={"ID":"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3","Type":"ContainerStarted","Data":"03de61365dc640f811469917238ca009d3aa1f975ae5e1adb612a656dc7a4b32"} Dec 03 09:16:29 crc kubenswrapper[4573]: I1203 09:16:29.249350 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" podStartSLOduration=2.020170312 podStartE2EDuration="2.249327403s" podCreationTimestamp="2025-12-03 09:16:27 +0000 UTC" firstStartedPulling="2025-12-03 09:16:28.367626615 +0000 UTC m=+2308.936005874" lastFinishedPulling="2025-12-03 09:16:28.596783696 +0000 UTC m=+2309.165162965" observedRunningTime="2025-12-03 09:16:29.248704576 +0000 UTC m=+2309.817083835" watchObservedRunningTime="2025-12-03 09:16:29.249327403 +0000 UTC m=+2309.817706662" Dec 03 09:16:37 crc kubenswrapper[4573]: I1203 09:16:37.301844 4573 generic.go:334] "Generic (PLEG): container finished" podID="9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3" containerID="ad8608f6aa3b9ebf9287e1a4f7358de86e308537533378b1062a4fd1fb35d986" exitCode=0 Dec 03 09:16:37 crc kubenswrapper[4573]: I1203 09:16:37.301938 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" event={"ID":"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3","Type":"ContainerDied","Data":"ad8608f6aa3b9ebf9287e1a4f7358de86e308537533378b1062a4fd1fb35d986"} Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.747599 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.769069 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx5t8\" (UniqueName: \"kubernetes.io/projected/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-kube-api-access-xx5t8\") pod \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.769243 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-ssh-key-openstack-edpm-ipam\") pod \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.769303 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-inventory-0\") pod \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\" (UID: \"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3\") " Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.780233 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-kube-api-access-xx5t8" (OuterVolumeSpecName: "kube-api-access-xx5t8") pod "9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3" (UID: "9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3"). InnerVolumeSpecName "kube-api-access-xx5t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.825104 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3" (UID: "9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.827416 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3" (UID: "9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.878242 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx5t8\" (UniqueName: \"kubernetes.io/projected/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-kube-api-access-xx5t8\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.878555 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:38 crc kubenswrapper[4573]: I1203 09:16:38.878567 4573 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.325024 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" event={"ID":"9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3","Type":"ContainerDied","Data":"03de61365dc640f811469917238ca009d3aa1f975ae5e1adb612a656dc7a4b32"} Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.325114 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03de61365dc640f811469917238ca009d3aa1f975ae5e1adb612a656dc7a4b32" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.325235 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-gc8nt" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.471848 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j"] Dec 03 09:16:39 crc kubenswrapper[4573]: E1203 09:16:39.472355 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3" containerName="ssh-known-hosts-edpm-deployment" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.472375 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3" containerName="ssh-known-hosts-edpm-deployment" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.472582 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3" containerName="ssh-known-hosts-edpm-deployment" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.473228 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.478641 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.479670 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.479782 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.479882 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.492584 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j"] Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.599134 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.599553 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.599973 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hhzb\" (UniqueName: \"kubernetes.io/projected/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-kube-api-access-5hhzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.702600 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hhzb\" (UniqueName: \"kubernetes.io/projected/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-kube-api-access-5hhzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.702762 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.702960 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.716264 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.716366 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.732830 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hhzb\" (UniqueName: \"kubernetes.io/projected/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-kube-api-access-5hhzb\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-qnv8j\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:39 crc kubenswrapper[4573]: I1203 09:16:39.792872 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:40 crc kubenswrapper[4573]: I1203 09:16:40.379858 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j"] Dec 03 09:16:41 crc kubenswrapper[4573]: I1203 09:16:41.348003 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" event={"ID":"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619","Type":"ContainerStarted","Data":"6df0d0c7f788b89640b589c77c9b05ca4c9ffd509bb915ade1cc046b56c142b1"} Dec 03 09:16:42 crc kubenswrapper[4573]: I1203 09:16:42.357902 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" event={"ID":"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619","Type":"ContainerStarted","Data":"ec458497df6efa28b0c2a3ca52782688a8ee3c427ff9510d5b3ccfbbe671bc3d"} Dec 03 09:16:42 crc kubenswrapper[4573]: I1203 09:16:42.379634 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" podStartSLOduration=2.173779501 podStartE2EDuration="3.379608178s" podCreationTimestamp="2025-12-03 09:16:39 +0000 UTC" firstStartedPulling="2025-12-03 09:16:40.389947753 +0000 UTC m=+2320.958327012" lastFinishedPulling="2025-12-03 09:16:41.59577643 +0000 UTC m=+2322.164155689" observedRunningTime="2025-12-03 09:16:42.374808567 +0000 UTC m=+2322.943187846" watchObservedRunningTime="2025-12-03 09:16:42.379608178 +0000 UTC m=+2322.947987437" Dec 03 09:16:51 crc kubenswrapper[4573]: I1203 09:16:51.447951 4573 generic.go:334] "Generic (PLEG): container finished" podID="fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619" containerID="ec458497df6efa28b0c2a3ca52782688a8ee3c427ff9510d5b3ccfbbe671bc3d" exitCode=0 Dec 03 09:16:51 crc kubenswrapper[4573]: I1203 09:16:51.448074 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" event={"ID":"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619","Type":"ContainerDied","Data":"ec458497df6efa28b0c2a3ca52782688a8ee3c427ff9510d5b3ccfbbe671bc3d"} Dec 03 09:16:52 crc kubenswrapper[4573]: I1203 09:16:52.975778 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.155531 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-inventory\") pod \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.155873 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hhzb\" (UniqueName: \"kubernetes.io/projected/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-kube-api-access-5hhzb\") pod \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.156476 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-ssh-key\") pod \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\" (UID: \"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619\") " Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.162782 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-kube-api-access-5hhzb" (OuterVolumeSpecName: "kube-api-access-5hhzb") pod "fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619" (UID: "fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619"). InnerVolumeSpecName "kube-api-access-5hhzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.191524 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-inventory" (OuterVolumeSpecName: "inventory") pod "fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619" (UID: "fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.193710 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619" (UID: "fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.259441 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hhzb\" (UniqueName: \"kubernetes.io/projected/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-kube-api-access-5hhzb\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.259506 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.259518 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.469847 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" event={"ID":"fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619","Type":"ContainerDied","Data":"6df0d0c7f788b89640b589c77c9b05ca4c9ffd509bb915ade1cc046b56c142b1"} Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.469904 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6df0d0c7f788b89640b589c77c9b05ca4c9ffd509bb915ade1cc046b56c142b1" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.469925 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-qnv8j" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.679901 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp"] Dec 03 09:16:53 crc kubenswrapper[4573]: E1203 09:16:53.685025 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.685079 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.685327 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.686279 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.690644 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.690981 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.692199 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.693077 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.697773 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp"] Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.774166 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n28vx\" (UniqueName: \"kubernetes.io/projected/fb9f04b9-8dd3-4776-8dec-c469a94759c2-kube-api-access-n28vx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.774234 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.774462 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.877825 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n28vx\" (UniqueName: \"kubernetes.io/projected/fb9f04b9-8dd3-4776-8dec-c469a94759c2-kube-api-access-n28vx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.877903 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.877937 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.889572 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.889714 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:53 crc kubenswrapper[4573]: I1203 09:16:53.901767 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n28vx\" (UniqueName: \"kubernetes.io/projected/fb9f04b9-8dd3-4776-8dec-c469a94759c2-kube-api-access-n28vx\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:54 crc kubenswrapper[4573]: I1203 09:16:54.065867 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:16:54 crc kubenswrapper[4573]: I1203 09:16:54.657136 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp"] Dec 03 09:16:55 crc kubenswrapper[4573]: I1203 09:16:55.495530 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" event={"ID":"fb9f04b9-8dd3-4776-8dec-c469a94759c2","Type":"ContainerStarted","Data":"36d409aa5ae6829cf789cdab00d5b74402a0b2d4cac721920fe7ffdf71948a6a"} Dec 03 09:16:55 crc kubenswrapper[4573]: I1203 09:16:55.495973 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" event={"ID":"fb9f04b9-8dd3-4776-8dec-c469a94759c2","Type":"ContainerStarted","Data":"34fb0f97c61a4e81ff6936ffe4f61761da3893149cf5b2894081263028b5bc62"} Dec 03 09:16:55 crc kubenswrapper[4573]: I1203 09:16:55.532132 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" podStartSLOduration=2.338401765 podStartE2EDuration="2.532110418s" podCreationTimestamp="2025-12-03 09:16:53 +0000 UTC" firstStartedPulling="2025-12-03 09:16:54.672160384 +0000 UTC m=+2335.240539643" lastFinishedPulling="2025-12-03 09:16:54.865869037 +0000 UTC m=+2335.434248296" observedRunningTime="2025-12-03 09:16:55.522131156 +0000 UTC m=+2336.090510415" watchObservedRunningTime="2025-12-03 09:16:55.532110418 +0000 UTC m=+2336.100489677" Dec 03 09:16:56 crc kubenswrapper[4573]: I1203 09:16:56.943717 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:16:56 crc kubenswrapper[4573]: I1203 09:16:56.945020 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:17:06 crc kubenswrapper[4573]: I1203 09:17:06.630884 4573 generic.go:334] "Generic (PLEG): container finished" podID="fb9f04b9-8dd3-4776-8dec-c469a94759c2" containerID="36d409aa5ae6829cf789cdab00d5b74402a0b2d4cac721920fe7ffdf71948a6a" exitCode=0 Dec 03 09:17:06 crc kubenswrapper[4573]: I1203 09:17:06.630981 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" event={"ID":"fb9f04b9-8dd3-4776-8dec-c469a94759c2","Type":"ContainerDied","Data":"36d409aa5ae6829cf789cdab00d5b74402a0b2d4cac721920fe7ffdf71948a6a"} Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.096896 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.257391 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-inventory\") pod \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.257451 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-ssh-key\") pod \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.257608 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n28vx\" (UniqueName: \"kubernetes.io/projected/fb9f04b9-8dd3-4776-8dec-c469a94759c2-kube-api-access-n28vx\") pod \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\" (UID: \"fb9f04b9-8dd3-4776-8dec-c469a94759c2\") " Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.268450 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb9f04b9-8dd3-4776-8dec-c469a94759c2-kube-api-access-n28vx" (OuterVolumeSpecName: "kube-api-access-n28vx") pod "fb9f04b9-8dd3-4776-8dec-c469a94759c2" (UID: "fb9f04b9-8dd3-4776-8dec-c469a94759c2"). InnerVolumeSpecName "kube-api-access-n28vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.295940 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb9f04b9-8dd3-4776-8dec-c469a94759c2" (UID: "fb9f04b9-8dd3-4776-8dec-c469a94759c2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.301547 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-inventory" (OuterVolumeSpecName: "inventory") pod "fb9f04b9-8dd3-4776-8dec-c469a94759c2" (UID: "fb9f04b9-8dd3-4776-8dec-c469a94759c2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.360149 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n28vx\" (UniqueName: \"kubernetes.io/projected/fb9f04b9-8dd3-4776-8dec-c469a94759c2-kube-api-access-n28vx\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.360187 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.360197 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb9f04b9-8dd3-4776-8dec-c469a94759c2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.657182 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" event={"ID":"fb9f04b9-8dd3-4776-8dec-c469a94759c2","Type":"ContainerDied","Data":"34fb0f97c61a4e81ff6936ffe4f61761da3893149cf5b2894081263028b5bc62"} Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.657245 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34fb0f97c61a4e81ff6936ffe4f61761da3893149cf5b2894081263028b5bc62" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.657298 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.777471 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p"] Dec 03 09:17:08 crc kubenswrapper[4573]: E1203 09:17:08.779655 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb9f04b9-8dd3-4776-8dec-c469a94759c2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.779697 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb9f04b9-8dd3-4776-8dec-c469a94759c2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.779995 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb9f04b9-8dd3-4776-8dec-c469a94759c2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.781068 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.789139 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.789303 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.789529 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.789659 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.789696 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.789875 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.790154 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.790981 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.803098 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p"] Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.877474 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.877838 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.877963 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878336 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878411 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878511 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878552 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878615 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878797 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878882 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878912 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.878959 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvc4h\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-kube-api-access-zvc4h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.879186 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.879241 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.981676 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.981770 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.981842 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.981876 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.981908 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982004 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982038 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982104 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982137 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982176 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982217 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982244 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982275 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.982303 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvc4h\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-kube-api-access-zvc4h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.987349 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.987691 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.989016 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.991782 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.992396 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.992485 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.992921 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.993158 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.993381 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.995578 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:08 crc kubenswrapper[4573]: I1203 09:17:08.995741 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:09 crc kubenswrapper[4573]: I1203 09:17:09.001144 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:09 crc kubenswrapper[4573]: I1203 09:17:09.001758 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:09 crc kubenswrapper[4573]: I1203 09:17:09.005082 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvc4h\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-kube-api-access-zvc4h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:09 crc kubenswrapper[4573]: I1203 09:17:09.113966 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:09 crc kubenswrapper[4573]: I1203 09:17:09.719619 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p"] Dec 03 09:17:10 crc kubenswrapper[4573]: I1203 09:17:10.680828 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" event={"ID":"66bd3982-db80-45dc-a985-5b2f05a0e257","Type":"ContainerStarted","Data":"ab08014d1e0c26b244a025bf34da2450ab2750957968b1ce7d58dbfe030421e0"} Dec 03 09:17:10 crc kubenswrapper[4573]: I1203 09:17:10.682092 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" event={"ID":"66bd3982-db80-45dc-a985-5b2f05a0e257","Type":"ContainerStarted","Data":"a6c2b9c9da82d8e2ecc35361a520ae51674f19a9e7c6e1eb75c9ea94445c064a"} Dec 03 09:17:10 crc kubenswrapper[4573]: I1203 09:17:10.721127 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" podStartSLOduration=2.542287535 podStartE2EDuration="2.721107583s" podCreationTimestamp="2025-12-03 09:17:08 +0000 UTC" firstStartedPulling="2025-12-03 09:17:09.720702318 +0000 UTC m=+2350.289081577" lastFinishedPulling="2025-12-03 09:17:09.899522366 +0000 UTC m=+2350.467901625" observedRunningTime="2025-12-03 09:17:10.714699529 +0000 UTC m=+2351.283078828" watchObservedRunningTime="2025-12-03 09:17:10.721107583 +0000 UTC m=+2351.289486842" Dec 03 09:17:26 crc kubenswrapper[4573]: I1203 09:17:26.944264 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:17:26 crc kubenswrapper[4573]: I1203 09:17:26.945993 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:17:54 crc kubenswrapper[4573]: I1203 09:17:54.150698 4573 generic.go:334] "Generic (PLEG): container finished" podID="66bd3982-db80-45dc-a985-5b2f05a0e257" containerID="ab08014d1e0c26b244a025bf34da2450ab2750957968b1ce7d58dbfe030421e0" exitCode=0 Dec 03 09:17:54 crc kubenswrapper[4573]: I1203 09:17:54.150794 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" event={"ID":"66bd3982-db80-45dc-a985-5b2f05a0e257","Type":"ContainerDied","Data":"ab08014d1e0c26b244a025bf34da2450ab2750957968b1ce7d58dbfe030421e0"} Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.745874 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856180 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-libvirt-combined-ca-bundle\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856263 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-neutron-metadata-combined-ca-bundle\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856301 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-ovn-default-certs-0\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856415 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-bootstrap-combined-ca-bundle\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856494 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-inventory\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856555 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856609 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856713 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-telemetry-combined-ca-bundle\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856739 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-repo-setup-combined-ca-bundle\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856806 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ovn-combined-ca-bundle\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856836 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856871 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvc4h\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-kube-api-access-zvc4h\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856897 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-nova-combined-ca-bundle\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.856927 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ssh-key\") pod \"66bd3982-db80-45dc-a985-5b2f05a0e257\" (UID: \"66bd3982-db80-45dc-a985-5b2f05a0e257\") " Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.867095 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.870207 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.870625 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.870667 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.871880 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.872529 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.879858 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.882196 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-kube-api-access-zvc4h" (OuterVolumeSpecName: "kube-api-access-zvc4h") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "kube-api-access-zvc4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.883017 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.883696 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.885606 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.896392 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.906690 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.916961 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-inventory" (OuterVolumeSpecName: "inventory") pod "66bd3982-db80-45dc-a985-5b2f05a0e257" (UID: "66bd3982-db80-45dc-a985-5b2f05a0e257"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959411 4573 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959473 4573 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959484 4573 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959501 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959512 4573 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959526 4573 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959539 4573 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959549 4573 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959558 4573 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959569 4573 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959578 4573 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959659 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvc4h\" (UniqueName: \"kubernetes.io/projected/66bd3982-db80-45dc-a985-5b2f05a0e257-kube-api-access-zvc4h\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959670 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:55 crc kubenswrapper[4573]: I1203 09:17:55.959679 4573 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66bd3982-db80-45dc-a985-5b2f05a0e257-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.171079 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" event={"ID":"66bd3982-db80-45dc-a985-5b2f05a0e257","Type":"ContainerDied","Data":"a6c2b9c9da82d8e2ecc35361a520ae51674f19a9e7c6e1eb75c9ea94445c064a"} Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.171135 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6c2b9c9da82d8e2ecc35361a520ae51674f19a9e7c6e1eb75c9ea94445c064a" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.171146 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.301849 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw"] Dec 03 09:17:56 crc kubenswrapper[4573]: E1203 09:17:56.302400 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66bd3982-db80-45dc-a985-5b2f05a0e257" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.302423 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="66bd3982-db80-45dc-a985-5b2f05a0e257" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.302628 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="66bd3982-db80-45dc-a985-5b2f05a0e257" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.303361 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.306418 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.306634 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.306860 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.307085 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.307284 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.320994 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw"] Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.470617 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.470690 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.471068 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.471323 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.471473 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ngh2\" (UniqueName: \"kubernetes.io/projected/633c66a3-1e25-4f81-9eee-2cb09bf54855-kube-api-access-4ngh2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.573620 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ngh2\" (UniqueName: \"kubernetes.io/projected/633c66a3-1e25-4f81-9eee-2cb09bf54855-kube-api-access-4ngh2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.573725 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.573755 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.573823 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.573901 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.575485 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.577780 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.578601 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.581418 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.595119 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ngh2\" (UniqueName: \"kubernetes.io/projected/633c66a3-1e25-4f81-9eee-2cb09bf54855-kube-api-access-4ngh2\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-vq5qw\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.626838 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.943369 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.943818 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.943869 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.944798 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:17:56 crc kubenswrapper[4573]: I1203 09:17:56.944867 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" gracePeriod=600 Dec 03 09:17:57 crc kubenswrapper[4573]: I1203 09:17:57.251451 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw"] Dec 03 09:17:57 crc kubenswrapper[4573]: E1203 09:17:57.615533 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:17:58 crc kubenswrapper[4573]: I1203 09:17:58.194645 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" exitCode=0 Dec 03 09:17:58 crc kubenswrapper[4573]: I1203 09:17:58.194726 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3"} Dec 03 09:17:58 crc kubenswrapper[4573]: I1203 09:17:58.194783 4573 scope.go:117] "RemoveContainer" containerID="300e60f96de70ba98f90e77595dce33621c17e38eac9f7c366d93366d0b0c462" Dec 03 09:17:58 crc kubenswrapper[4573]: I1203 09:17:58.196091 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:17:58 crc kubenswrapper[4573]: E1203 09:17:58.196737 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:17:58 crc kubenswrapper[4573]: I1203 09:17:58.200166 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" event={"ID":"633c66a3-1e25-4f81-9eee-2cb09bf54855","Type":"ContainerStarted","Data":"d749cd76dce3a1987f2def120a3f6060f9c031983ea32b4df7829da2f6666460"} Dec 03 09:17:58 crc kubenswrapper[4573]: I1203 09:17:58.200246 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" event={"ID":"633c66a3-1e25-4f81-9eee-2cb09bf54855","Type":"ContainerStarted","Data":"afce1308274e8ce5a5f5d15264447f0f0edfa76134e466206e416d0a0502b3aa"} Dec 03 09:18:09 crc kubenswrapper[4573]: I1203 09:18:09.031118 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:18:09 crc kubenswrapper[4573]: E1203 09:18:09.032453 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:18:24 crc kubenswrapper[4573]: I1203 09:18:24.030469 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:18:24 crc kubenswrapper[4573]: E1203 09:18:24.031281 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:18:36 crc kubenswrapper[4573]: I1203 09:18:36.030568 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:18:36 crc kubenswrapper[4573]: E1203 09:18:36.031666 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:18:47 crc kubenswrapper[4573]: I1203 09:18:47.030612 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:18:47 crc kubenswrapper[4573]: E1203 09:18:47.031481 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:18:59 crc kubenswrapper[4573]: I1203 09:18:59.029839 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:18:59 crc kubenswrapper[4573]: E1203 09:18:59.030562 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:19:11 crc kubenswrapper[4573]: I1203 09:19:11.055983 4573 generic.go:334] "Generic (PLEG): container finished" podID="633c66a3-1e25-4f81-9eee-2cb09bf54855" containerID="d749cd76dce3a1987f2def120a3f6060f9c031983ea32b4df7829da2f6666460" exitCode=0 Dec 03 09:19:11 crc kubenswrapper[4573]: I1203 09:19:11.056135 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" event={"ID":"633c66a3-1e25-4f81-9eee-2cb09bf54855","Type":"ContainerDied","Data":"d749cd76dce3a1987f2def120a3f6060f9c031983ea32b4df7829da2f6666460"} Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.518871 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.699419 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-inventory\") pod \"633c66a3-1e25-4f81-9eee-2cb09bf54855\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.699480 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ssh-key\") pod \"633c66a3-1e25-4f81-9eee-2cb09bf54855\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.699510 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovn-combined-ca-bundle\") pod \"633c66a3-1e25-4f81-9eee-2cb09bf54855\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.699557 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovncontroller-config-0\") pod \"633c66a3-1e25-4f81-9eee-2cb09bf54855\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.699680 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ngh2\" (UniqueName: \"kubernetes.io/projected/633c66a3-1e25-4f81-9eee-2cb09bf54855-kube-api-access-4ngh2\") pod \"633c66a3-1e25-4f81-9eee-2cb09bf54855\" (UID: \"633c66a3-1e25-4f81-9eee-2cb09bf54855\") " Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.708235 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/633c66a3-1e25-4f81-9eee-2cb09bf54855-kube-api-access-4ngh2" (OuterVolumeSpecName: "kube-api-access-4ngh2") pod "633c66a3-1e25-4f81-9eee-2cb09bf54855" (UID: "633c66a3-1e25-4f81-9eee-2cb09bf54855"). InnerVolumeSpecName "kube-api-access-4ngh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.708601 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "633c66a3-1e25-4f81-9eee-2cb09bf54855" (UID: "633c66a3-1e25-4f81-9eee-2cb09bf54855"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.733869 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "633c66a3-1e25-4f81-9eee-2cb09bf54855" (UID: "633c66a3-1e25-4f81-9eee-2cb09bf54855"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.734813 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "633c66a3-1e25-4f81-9eee-2cb09bf54855" (UID: "633c66a3-1e25-4f81-9eee-2cb09bf54855"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.748998 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-inventory" (OuterVolumeSpecName: "inventory") pod "633c66a3-1e25-4f81-9eee-2cb09bf54855" (UID: "633c66a3-1e25-4f81-9eee-2cb09bf54855"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.802678 4573 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.802745 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ngh2\" (UniqueName: \"kubernetes.io/projected/633c66a3-1e25-4f81-9eee-2cb09bf54855-kube-api-access-4ngh2\") on node \"crc\" DevicePath \"\"" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.802756 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.802766 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:19:12 crc kubenswrapper[4573]: I1203 09:19:12.802778 4573 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/633c66a3-1e25-4f81-9eee-2cb09bf54855-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.085172 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" event={"ID":"633c66a3-1e25-4f81-9eee-2cb09bf54855","Type":"ContainerDied","Data":"afce1308274e8ce5a5f5d15264447f0f0edfa76134e466206e416d0a0502b3aa"} Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.085228 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afce1308274e8ce5a5f5d15264447f0f0edfa76134e466206e416d0a0502b3aa" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.085335 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-vq5qw" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.237434 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44"] Dec 03 09:19:13 crc kubenswrapper[4573]: E1203 09:19:13.240097 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633c66a3-1e25-4f81-9eee-2cb09bf54855" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.240139 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="633c66a3-1e25-4f81-9eee-2cb09bf54855" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.240418 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="633c66a3-1e25-4f81-9eee-2cb09bf54855" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.241397 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.266395 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.266469 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.266609 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.266616 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.266700 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.267185 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.271009 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44"] Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.422474 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.422655 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.422837 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8c98\" (UniqueName: \"kubernetes.io/projected/281980d3-9cd5-4f71-b6a1-8602818bc836-kube-api-access-n8c98\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.422898 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.422931 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.423002 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.525793 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.525914 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.526033 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.526245 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8c98\" (UniqueName: \"kubernetes.io/projected/281980d3-9cd5-4f71-b6a1-8602818bc836-kube-api-access-n8c98\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.526296 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.526328 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.532583 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.533329 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.533635 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.533948 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.543294 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.555803 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8c98\" (UniqueName: \"kubernetes.io/projected/281980d3-9cd5-4f71-b6a1-8602818bc836-kube-api-access-n8c98\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:13 crc kubenswrapper[4573]: I1203 09:19:13.563618 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:19:14 crc kubenswrapper[4573]: I1203 09:19:14.031225 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:19:14 crc kubenswrapper[4573]: E1203 09:19:14.031820 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:19:14 crc kubenswrapper[4573]: I1203 09:19:14.167702 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44"] Dec 03 09:19:15 crc kubenswrapper[4573]: I1203 09:19:15.106750 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" event={"ID":"281980d3-9cd5-4f71-b6a1-8602818bc836","Type":"ContainerStarted","Data":"111898d6f9201f5a194b3192f973bae2f4bc9724167c091e67542537f4908119"} Dec 03 09:19:15 crc kubenswrapper[4573]: I1203 09:19:15.107219 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" event={"ID":"281980d3-9cd5-4f71-b6a1-8602818bc836","Type":"ContainerStarted","Data":"78df1712fd5da7472552323ecf7151ea2cb8cbebe7e0bb7dbde9d157e3484605"} Dec 03 09:19:15 crc kubenswrapper[4573]: I1203 09:19:15.156032 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" podStartSLOduration=2.006407597 podStartE2EDuration="2.156002016s" podCreationTimestamp="2025-12-03 09:19:13 +0000 UTC" firstStartedPulling="2025-12-03 09:19:14.172218796 +0000 UTC m=+2474.740598055" lastFinishedPulling="2025-12-03 09:19:14.321813205 +0000 UTC m=+2474.890192474" observedRunningTime="2025-12-03 09:19:15.142696963 +0000 UTC m=+2475.711076262" watchObservedRunningTime="2025-12-03 09:19:15.156002016 +0000 UTC m=+2475.724381295" Dec 03 09:19:25 crc kubenswrapper[4573]: I1203 09:19:25.031178 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:19:25 crc kubenswrapper[4573]: E1203 09:19:25.031875 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:19:39 crc kubenswrapper[4573]: I1203 09:19:39.030728 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:19:39 crc kubenswrapper[4573]: E1203 09:19:39.031793 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:19:50 crc kubenswrapper[4573]: I1203 09:19:50.037463 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:19:50 crc kubenswrapper[4573]: E1203 09:19:50.038482 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:20:02 crc kubenswrapper[4573]: I1203 09:20:02.031166 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:20:02 crc kubenswrapper[4573]: E1203 09:20:02.032374 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:20:08 crc kubenswrapper[4573]: I1203 09:20:08.672605 4573 generic.go:334] "Generic (PLEG): container finished" podID="281980d3-9cd5-4f71-b6a1-8602818bc836" containerID="111898d6f9201f5a194b3192f973bae2f4bc9724167c091e67542537f4908119" exitCode=0 Dec 03 09:20:08 crc kubenswrapper[4573]: I1203 09:20:08.672783 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" event={"ID":"281980d3-9cd5-4f71-b6a1-8602818bc836","Type":"ContainerDied","Data":"111898d6f9201f5a194b3192f973bae2f4bc9724167c091e67542537f4908119"} Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.164214 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.321406 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-metadata-combined-ca-bundle\") pod \"281980d3-9cd5-4f71-b6a1-8602818bc836\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.321518 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-ssh-key\") pod \"281980d3-9cd5-4f71-b6a1-8602818bc836\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.321579 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8c98\" (UniqueName: \"kubernetes.io/projected/281980d3-9cd5-4f71-b6a1-8602818bc836-kube-api-access-n8c98\") pod \"281980d3-9cd5-4f71-b6a1-8602818bc836\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.321742 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-inventory\") pod \"281980d3-9cd5-4f71-b6a1-8602818bc836\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.321819 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-ovn-metadata-agent-neutron-config-0\") pod \"281980d3-9cd5-4f71-b6a1-8602818bc836\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.321886 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-nova-metadata-neutron-config-0\") pod \"281980d3-9cd5-4f71-b6a1-8602818bc836\" (UID: \"281980d3-9cd5-4f71-b6a1-8602818bc836\") " Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.328589 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "281980d3-9cd5-4f71-b6a1-8602818bc836" (UID: "281980d3-9cd5-4f71-b6a1-8602818bc836"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.328912 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/281980d3-9cd5-4f71-b6a1-8602818bc836-kube-api-access-n8c98" (OuterVolumeSpecName: "kube-api-access-n8c98") pod "281980d3-9cd5-4f71-b6a1-8602818bc836" (UID: "281980d3-9cd5-4f71-b6a1-8602818bc836"). InnerVolumeSpecName "kube-api-access-n8c98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.354741 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "281980d3-9cd5-4f71-b6a1-8602818bc836" (UID: "281980d3-9cd5-4f71-b6a1-8602818bc836"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.354869 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "281980d3-9cd5-4f71-b6a1-8602818bc836" (UID: "281980d3-9cd5-4f71-b6a1-8602818bc836"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.364009 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "281980d3-9cd5-4f71-b6a1-8602818bc836" (UID: "281980d3-9cd5-4f71-b6a1-8602818bc836"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.364957 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-inventory" (OuterVolumeSpecName: "inventory") pod "281980d3-9cd5-4f71-b6a1-8602818bc836" (UID: "281980d3-9cd5-4f71-b6a1-8602818bc836"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.427731 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.427783 4573 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.427798 4573 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.427815 4573 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.427831 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/281980d3-9cd5-4f71-b6a1-8602818bc836-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.427844 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8c98\" (UniqueName: \"kubernetes.io/projected/281980d3-9cd5-4f71-b6a1-8602818bc836-kube-api-access-n8c98\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.699281 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" event={"ID":"281980d3-9cd5-4f71-b6a1-8602818bc836","Type":"ContainerDied","Data":"78df1712fd5da7472552323ecf7151ea2cb8cbebe7e0bb7dbde9d157e3484605"} Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.699342 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78df1712fd5da7472552323ecf7151ea2cb8cbebe7e0bb7dbde9d157e3484605" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.699436 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.817744 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d"] Dec 03 09:20:10 crc kubenswrapper[4573]: E1203 09:20:10.818232 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="281980d3-9cd5-4f71-b6a1-8602818bc836" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.818250 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="281980d3-9cd5-4f71-b6a1-8602818bc836" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.818439 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="281980d3-9cd5-4f71-b6a1-8602818bc836" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.819155 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.824526 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.824740 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.824849 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.825218 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.825315 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.879781 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d"] Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.939412 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.939474 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.939526 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.939592 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dvwf\" (UniqueName: \"kubernetes.io/projected/022b0b02-d8bb-4560-af50-5d048f4ec520-kube-api-access-8dvwf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:10 crc kubenswrapper[4573]: I1203 09:20:10.939633 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: E1203 09:20:11.016634 4573 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod281980d3_9cd5_4f71_b6a1_8602818bc836.slice/crio-78df1712fd5da7472552323ecf7151ea2cb8cbebe7e0bb7dbde9d157e3484605\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod281980d3_9cd5_4f71_b6a1_8602818bc836.slice\": RecentStats: unable to find data in memory cache]" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.040694 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dvwf\" (UniqueName: \"kubernetes.io/projected/022b0b02-d8bb-4560-af50-5d048f4ec520-kube-api-access-8dvwf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.040907 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.040988 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.041016 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.041068 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.046576 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.046830 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.055601 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.064756 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.077617 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dvwf\" (UniqueName: \"kubernetes.io/projected/022b0b02-d8bb-4560-af50-5d048f4ec520-kube-api-access-8dvwf\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.197737 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:20:11 crc kubenswrapper[4573]: I1203 09:20:11.775609 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d"] Dec 03 09:20:11 crc kubenswrapper[4573]: W1203 09:20:11.781364 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod022b0b02_d8bb_4560_af50_5d048f4ec520.slice/crio-9a2875012ff3ba7a96456da2c61ff54f954c808499c625db82aba7c4dfed9b58 WatchSource:0}: Error finding container 9a2875012ff3ba7a96456da2c61ff54f954c808499c625db82aba7c4dfed9b58: Status 404 returned error can't find the container with id 9a2875012ff3ba7a96456da2c61ff54f954c808499c625db82aba7c4dfed9b58 Dec 03 09:20:12 crc kubenswrapper[4573]: I1203 09:20:12.720261 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" event={"ID":"022b0b02-d8bb-4560-af50-5d048f4ec520","Type":"ContainerStarted","Data":"c3d2fd0746d561814dcb592b116d16cd00fe274dcc412a309947f9616e8d1135"} Dec 03 09:20:12 crc kubenswrapper[4573]: I1203 09:20:12.720684 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" event={"ID":"022b0b02-d8bb-4560-af50-5d048f4ec520","Type":"ContainerStarted","Data":"9a2875012ff3ba7a96456da2c61ff54f954c808499c625db82aba7c4dfed9b58"} Dec 03 09:20:12 crc kubenswrapper[4573]: I1203 09:20:12.745259 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" podStartSLOduration=2.55258767 podStartE2EDuration="2.745238474s" podCreationTimestamp="2025-12-03 09:20:10 +0000 UTC" firstStartedPulling="2025-12-03 09:20:11.784778769 +0000 UTC m=+2532.353158028" lastFinishedPulling="2025-12-03 09:20:11.977429573 +0000 UTC m=+2532.545808832" observedRunningTime="2025-12-03 09:20:12.737619057 +0000 UTC m=+2533.305998316" watchObservedRunningTime="2025-12-03 09:20:12.745238474 +0000 UTC m=+2533.313617733" Dec 03 09:20:15 crc kubenswrapper[4573]: I1203 09:20:15.031381 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:20:15 crc kubenswrapper[4573]: E1203 09:20:15.031947 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:20:29 crc kubenswrapper[4573]: I1203 09:20:29.031411 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:20:29 crc kubenswrapper[4573]: E1203 09:20:29.032667 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.260843 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwlj"] Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.264192 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.284593 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwlj"] Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.363248 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-utilities\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.363367 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw7r9\" (UniqueName: \"kubernetes.io/projected/f1558ad3-c1c5-4760-81f4-40040b79096f-kube-api-access-fw7r9\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.363722 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-catalog-content\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.466960 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw7r9\" (UniqueName: \"kubernetes.io/projected/f1558ad3-c1c5-4760-81f4-40040b79096f-kube-api-access-fw7r9\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.467090 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-catalog-content\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.467246 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-utilities\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.467860 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-catalog-content\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.467896 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-utilities\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.494394 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw7r9\" (UniqueName: \"kubernetes.io/projected/f1558ad3-c1c5-4760-81f4-40040b79096f-kube-api-access-fw7r9\") pod \"redhat-marketplace-mnwlj\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:36 crc kubenswrapper[4573]: I1203 09:20:36.593684 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:37 crc kubenswrapper[4573]: I1203 09:20:37.266867 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwlj"] Dec 03 09:20:38 crc kubenswrapper[4573]: I1203 09:20:38.004665 4573 generic.go:334] "Generic (PLEG): container finished" podID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerID="26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46" exitCode=0 Dec 03 09:20:38 crc kubenswrapper[4573]: I1203 09:20:38.004779 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwlj" event={"ID":"f1558ad3-c1c5-4760-81f4-40040b79096f","Type":"ContainerDied","Data":"26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46"} Dec 03 09:20:38 crc kubenswrapper[4573]: I1203 09:20:38.005069 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwlj" event={"ID":"f1558ad3-c1c5-4760-81f4-40040b79096f","Type":"ContainerStarted","Data":"69ee3f110f24255166af8c5a8120f969ebdbcaebd721b0a8c44d8e8343a35c9c"} Dec 03 09:20:38 crc kubenswrapper[4573]: I1203 09:20:38.009319 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:20:40 crc kubenswrapper[4573]: I1203 09:20:40.035898 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:20:40 crc kubenswrapper[4573]: E1203 09:20:40.036652 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:20:40 crc kubenswrapper[4573]: I1203 09:20:40.044609 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwlj" event={"ID":"f1558ad3-c1c5-4760-81f4-40040b79096f","Type":"ContainerStarted","Data":"b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e"} Dec 03 09:20:41 crc kubenswrapper[4573]: I1203 09:20:41.053413 4573 generic.go:334] "Generic (PLEG): container finished" podID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerID="b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e" exitCode=0 Dec 03 09:20:41 crc kubenswrapper[4573]: I1203 09:20:41.053467 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwlj" event={"ID":"f1558ad3-c1c5-4760-81f4-40040b79096f","Type":"ContainerDied","Data":"b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e"} Dec 03 09:20:42 crc kubenswrapper[4573]: I1203 09:20:42.071972 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwlj" event={"ID":"f1558ad3-c1c5-4760-81f4-40040b79096f","Type":"ContainerStarted","Data":"a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639"} Dec 03 09:20:42 crc kubenswrapper[4573]: I1203 09:20:42.100406 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mnwlj" podStartSLOduration=2.579988921 podStartE2EDuration="6.100386555s" podCreationTimestamp="2025-12-03 09:20:36 +0000 UTC" firstStartedPulling="2025-12-03 09:20:38.008901735 +0000 UTC m=+2558.577280994" lastFinishedPulling="2025-12-03 09:20:41.529299369 +0000 UTC m=+2562.097678628" observedRunningTime="2025-12-03 09:20:42.094411563 +0000 UTC m=+2562.662790832" watchObservedRunningTime="2025-12-03 09:20:42.100386555 +0000 UTC m=+2562.668765814" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.026458 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bmcs4"] Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.028775 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.064885 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bmcs4"] Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.163399 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-utilities\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.163726 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-catalog-content\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.164072 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwj8r\" (UniqueName: \"kubernetes.io/projected/85bba5be-441a-4f77-9a01-5a191e13ab6c-kube-api-access-rwj8r\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.267927 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-catalog-content\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.268028 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwj8r\" (UniqueName: \"kubernetes.io/projected/85bba5be-441a-4f77-9a01-5a191e13ab6c-kube-api-access-rwj8r\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.268137 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-utilities\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.268692 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-utilities\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.268973 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-catalog-content\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.287384 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwj8r\" (UniqueName: \"kubernetes.io/projected/85bba5be-441a-4f77-9a01-5a191e13ab6c-kube-api-access-rwj8r\") pod \"community-operators-bmcs4\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:43 crc kubenswrapper[4573]: I1203 09:20:43.356926 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:44 crc kubenswrapper[4573]: I1203 09:20:44.063297 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bmcs4"] Dec 03 09:20:44 crc kubenswrapper[4573]: I1203 09:20:44.108351 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmcs4" event={"ID":"85bba5be-441a-4f77-9a01-5a191e13ab6c","Type":"ContainerStarted","Data":"d78dd7228b8be38d62503571996fe4b35f15fb5825636e3b4e202cef967a48a1"} Dec 03 09:20:45 crc kubenswrapper[4573]: I1203 09:20:45.124695 4573 generic.go:334] "Generic (PLEG): container finished" podID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerID="a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994" exitCode=0 Dec 03 09:20:45 crc kubenswrapper[4573]: I1203 09:20:45.124806 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmcs4" event={"ID":"85bba5be-441a-4f77-9a01-5a191e13ab6c","Type":"ContainerDied","Data":"a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994"} Dec 03 09:20:46 crc kubenswrapper[4573]: I1203 09:20:46.594482 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:46 crc kubenswrapper[4573]: I1203 09:20:46.594952 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:46 crc kubenswrapper[4573]: I1203 09:20:46.649991 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:47 crc kubenswrapper[4573]: I1203 09:20:47.153818 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmcs4" event={"ID":"85bba5be-441a-4f77-9a01-5a191e13ab6c","Type":"ContainerStarted","Data":"55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5"} Dec 03 09:20:47 crc kubenswrapper[4573]: I1203 09:20:47.224803 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:48 crc kubenswrapper[4573]: I1203 09:20:48.421413 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwlj"] Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.178018 4573 generic.go:334] "Generic (PLEG): container finished" podID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerID="55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5" exitCode=0 Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.178585 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mnwlj" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerName="registry-server" containerID="cri-o://a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639" gracePeriod=2 Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.178155 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmcs4" event={"ID":"85bba5be-441a-4f77-9a01-5a191e13ab6c","Type":"ContainerDied","Data":"55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5"} Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.639405 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.717460 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw7r9\" (UniqueName: \"kubernetes.io/projected/f1558ad3-c1c5-4760-81f4-40040b79096f-kube-api-access-fw7r9\") pod \"f1558ad3-c1c5-4760-81f4-40040b79096f\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.721466 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-utilities\") pod \"f1558ad3-c1c5-4760-81f4-40040b79096f\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.721532 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-catalog-content\") pod \"f1558ad3-c1c5-4760-81f4-40040b79096f\" (UID: \"f1558ad3-c1c5-4760-81f4-40040b79096f\") " Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.723194 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-utilities" (OuterVolumeSpecName: "utilities") pod "f1558ad3-c1c5-4760-81f4-40040b79096f" (UID: "f1558ad3-c1c5-4760-81f4-40040b79096f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.723767 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1558ad3-c1c5-4760-81f4-40040b79096f-kube-api-access-fw7r9" (OuterVolumeSpecName: "kube-api-access-fw7r9") pod "f1558ad3-c1c5-4760-81f4-40040b79096f" (UID: "f1558ad3-c1c5-4760-81f4-40040b79096f"). InnerVolumeSpecName "kube-api-access-fw7r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.746288 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1558ad3-c1c5-4760-81f4-40040b79096f" (UID: "f1558ad3-c1c5-4760-81f4-40040b79096f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.824535 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.824574 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1558ad3-c1c5-4760-81f4-40040b79096f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:49 crc kubenswrapper[4573]: I1203 09:20:49.824586 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw7r9\" (UniqueName: \"kubernetes.io/projected/f1558ad3-c1c5-4760-81f4-40040b79096f-kube-api-access-fw7r9\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.211936 4573 generic.go:334] "Generic (PLEG): container finished" podID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerID="a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639" exitCode=0 Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.212374 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwlj" event={"ID":"f1558ad3-c1c5-4760-81f4-40040b79096f","Type":"ContainerDied","Data":"a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639"} Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.212417 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mnwlj" event={"ID":"f1558ad3-c1c5-4760-81f4-40040b79096f","Type":"ContainerDied","Data":"69ee3f110f24255166af8c5a8120f969ebdbcaebd721b0a8c44d8e8343a35c9c"} Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.212631 4573 scope.go:117] "RemoveContainer" containerID="a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.212876 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mnwlj" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.222535 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmcs4" event={"ID":"85bba5be-441a-4f77-9a01-5a191e13ab6c","Type":"ContainerStarted","Data":"b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb"} Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.248090 4573 scope.go:117] "RemoveContainer" containerID="b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.251923 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwlj"] Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.266256 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mnwlj"] Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.274439 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bmcs4" podStartSLOduration=2.727664646 podStartE2EDuration="7.274415424s" podCreationTimestamp="2025-12-03 09:20:43 +0000 UTC" firstStartedPulling="2025-12-03 09:20:45.128190586 +0000 UTC m=+2565.696569855" lastFinishedPulling="2025-12-03 09:20:49.674941374 +0000 UTC m=+2570.243320633" observedRunningTime="2025-12-03 09:20:50.264251776 +0000 UTC m=+2570.832631035" watchObservedRunningTime="2025-12-03 09:20:50.274415424 +0000 UTC m=+2570.842794683" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.280192 4573 scope.go:117] "RemoveContainer" containerID="26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.312300 4573 scope.go:117] "RemoveContainer" containerID="a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639" Dec 03 09:20:50 crc kubenswrapper[4573]: E1203 09:20:50.313007 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639\": container with ID starting with a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639 not found: ID does not exist" containerID="a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.313170 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639"} err="failed to get container status \"a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639\": rpc error: code = NotFound desc = could not find container \"a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639\": container with ID starting with a1f60fb86a7fb6aaec58989cd9705499541e12292b58afd910db9e0e91516639 not found: ID does not exist" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.313203 4573 scope.go:117] "RemoveContainer" containerID="b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e" Dec 03 09:20:50 crc kubenswrapper[4573]: E1203 09:20:50.313610 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e\": container with ID starting with b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e not found: ID does not exist" containerID="b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.313704 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e"} err="failed to get container status \"b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e\": rpc error: code = NotFound desc = could not find container \"b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e\": container with ID starting with b99d0052e55f1faa8b54599f5fb24e4dd6283524a1ca5457f6d69610fcdf155e not found: ID does not exist" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.313799 4573 scope.go:117] "RemoveContainer" containerID="26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46" Dec 03 09:20:50 crc kubenswrapper[4573]: E1203 09:20:50.314279 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46\": container with ID starting with 26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46 not found: ID does not exist" containerID="26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46" Dec 03 09:20:50 crc kubenswrapper[4573]: I1203 09:20:50.314309 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46"} err="failed to get container status \"26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46\": rpc error: code = NotFound desc = could not find container \"26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46\": container with ID starting with 26beadc902daf9aa28f842da5732c8a709d4e90430c461d2408879e8999c4e46 not found: ID does not exist" Dec 03 09:20:52 crc kubenswrapper[4573]: I1203 09:20:52.043772 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" path="/var/lib/kubelet/pods/f1558ad3-c1c5-4760-81f4-40040b79096f/volumes" Dec 03 09:20:53 crc kubenswrapper[4573]: I1203 09:20:53.358332 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:53 crc kubenswrapper[4573]: I1203 09:20:53.360544 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:53 crc kubenswrapper[4573]: I1203 09:20:53.433289 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:54 crc kubenswrapper[4573]: I1203 09:20:54.317187 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:55 crc kubenswrapper[4573]: I1203 09:20:55.019941 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bmcs4"] Dec 03 09:20:55 crc kubenswrapper[4573]: I1203 09:20:55.030034 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:20:55 crc kubenswrapper[4573]: E1203 09:20:55.030365 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.302405 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bmcs4" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerName="registry-server" containerID="cri-o://b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb" gracePeriod=2 Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.825852 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.886166 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-utilities\") pod \"85bba5be-441a-4f77-9a01-5a191e13ab6c\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.886315 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwj8r\" (UniqueName: \"kubernetes.io/projected/85bba5be-441a-4f77-9a01-5a191e13ab6c-kube-api-access-rwj8r\") pod \"85bba5be-441a-4f77-9a01-5a191e13ab6c\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.886374 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-catalog-content\") pod \"85bba5be-441a-4f77-9a01-5a191e13ab6c\" (UID: \"85bba5be-441a-4f77-9a01-5a191e13ab6c\") " Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.887289 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-utilities" (OuterVolumeSpecName: "utilities") pod "85bba5be-441a-4f77-9a01-5a191e13ab6c" (UID: "85bba5be-441a-4f77-9a01-5a191e13ab6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.901364 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85bba5be-441a-4f77-9a01-5a191e13ab6c-kube-api-access-rwj8r" (OuterVolumeSpecName: "kube-api-access-rwj8r") pod "85bba5be-441a-4f77-9a01-5a191e13ab6c" (UID: "85bba5be-441a-4f77-9a01-5a191e13ab6c"). InnerVolumeSpecName "kube-api-access-rwj8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.966400 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85bba5be-441a-4f77-9a01-5a191e13ab6c" (UID: "85bba5be-441a-4f77-9a01-5a191e13ab6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.988866 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwj8r\" (UniqueName: \"kubernetes.io/projected/85bba5be-441a-4f77-9a01-5a191e13ab6c-kube-api-access-rwj8r\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.988914 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:56 crc kubenswrapper[4573]: I1203 09:20:56.988926 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85bba5be-441a-4f77-9a01-5a191e13ab6c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.322106 4573 generic.go:334] "Generic (PLEG): container finished" podID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerID="b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb" exitCode=0 Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.322167 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmcs4" event={"ID":"85bba5be-441a-4f77-9a01-5a191e13ab6c","Type":"ContainerDied","Data":"b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb"} Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.322187 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bmcs4" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.322203 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmcs4" event={"ID":"85bba5be-441a-4f77-9a01-5a191e13ab6c","Type":"ContainerDied","Data":"d78dd7228b8be38d62503571996fe4b35f15fb5825636e3b4e202cef967a48a1"} Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.322228 4573 scope.go:117] "RemoveContainer" containerID="b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.345898 4573 scope.go:117] "RemoveContainer" containerID="55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.364584 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bmcs4"] Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.375511 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bmcs4"] Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.393842 4573 scope.go:117] "RemoveContainer" containerID="a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.437452 4573 scope.go:117] "RemoveContainer" containerID="b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb" Dec 03 09:20:57 crc kubenswrapper[4573]: E1203 09:20:57.438595 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb\": container with ID starting with b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb not found: ID does not exist" containerID="b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.438639 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb"} err="failed to get container status \"b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb\": rpc error: code = NotFound desc = could not find container \"b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb\": container with ID starting with b445d1ed31906fd631f454cdc7caf4c9e1bbb494c502ddadcb46f944c32b7adb not found: ID does not exist" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.438670 4573 scope.go:117] "RemoveContainer" containerID="55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5" Dec 03 09:20:57 crc kubenswrapper[4573]: E1203 09:20:57.439941 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5\": container with ID starting with 55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5 not found: ID does not exist" containerID="55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.439973 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5"} err="failed to get container status \"55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5\": rpc error: code = NotFound desc = could not find container \"55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5\": container with ID starting with 55ace596c50106dc8d7d9c4056c15d7527b490a88dcec95c1cc81340d8fe5ad5 not found: ID does not exist" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.439990 4573 scope.go:117] "RemoveContainer" containerID="a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994" Dec 03 09:20:57 crc kubenswrapper[4573]: E1203 09:20:57.440574 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994\": container with ID starting with a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994 not found: ID does not exist" containerID="a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994" Dec 03 09:20:57 crc kubenswrapper[4573]: I1203 09:20:57.440597 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994"} err="failed to get container status \"a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994\": rpc error: code = NotFound desc = could not find container \"a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994\": container with ID starting with a3e33d23b1f4ab929971e21ddee3c7503be60264d989ff3a515e885cb34cd994 not found: ID does not exist" Dec 03 09:20:58 crc kubenswrapper[4573]: I1203 09:20:58.056679 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" path="/var/lib/kubelet/pods/85bba5be-441a-4f77-9a01-5a191e13ab6c/volumes" Dec 03 09:21:07 crc kubenswrapper[4573]: I1203 09:21:07.030909 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:21:07 crc kubenswrapper[4573]: E1203 09:21:07.031774 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:21:20 crc kubenswrapper[4573]: I1203 09:21:20.038684 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:21:20 crc kubenswrapper[4573]: E1203 09:21:20.039806 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:21:34 crc kubenswrapper[4573]: I1203 09:21:34.031609 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:21:34 crc kubenswrapper[4573]: E1203 09:21:34.032815 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:21:47 crc kubenswrapper[4573]: I1203 09:21:47.031092 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:21:47 crc kubenswrapper[4573]: E1203 09:21:47.032169 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:22:02 crc kubenswrapper[4573]: I1203 09:22:02.030807 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:22:02 crc kubenswrapper[4573]: E1203 09:22:02.031769 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:22:14 crc kubenswrapper[4573]: I1203 09:22:14.030758 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:22:14 crc kubenswrapper[4573]: E1203 09:22:14.031739 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:22:29 crc kubenswrapper[4573]: I1203 09:22:29.031082 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:22:29 crc kubenswrapper[4573]: E1203 09:22:29.032439 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:22:40 crc kubenswrapper[4573]: I1203 09:22:40.039753 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:22:40 crc kubenswrapper[4573]: E1203 09:22:40.041033 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:22:54 crc kubenswrapper[4573]: I1203 09:22:54.030933 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:22:54 crc kubenswrapper[4573]: E1203 09:22:54.033435 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:23:07 crc kubenswrapper[4573]: I1203 09:23:07.032281 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:23:08 crc kubenswrapper[4573]: I1203 09:23:08.128237 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"9661b8f523a73ea76d82ecaf1583ed165c00467787ee991495f624ee4ae4fbd1"} Dec 03 09:23:22 crc kubenswrapper[4573]: I1203 09:23:22.418289 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg" podUID="a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:24:52 crc kubenswrapper[4573]: I1203 09:24:52.297997 4573 generic.go:334] "Generic (PLEG): container finished" podID="022b0b02-d8bb-4560-af50-5d048f4ec520" containerID="c3d2fd0746d561814dcb592b116d16cd00fe274dcc412a309947f9616e8d1135" exitCode=0 Dec 03 09:24:52 crc kubenswrapper[4573]: I1203 09:24:52.298222 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" event={"ID":"022b0b02-d8bb-4560-af50-5d048f4ec520","Type":"ContainerDied","Data":"c3d2fd0746d561814dcb592b116d16cd00fe274dcc412a309947f9616e8d1135"} Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.801175 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.807454 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dvwf\" (UniqueName: \"kubernetes.io/projected/022b0b02-d8bb-4560-af50-5d048f4ec520-kube-api-access-8dvwf\") pod \"022b0b02-d8bb-4560-af50-5d048f4ec520\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.807566 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-inventory\") pod \"022b0b02-d8bb-4560-af50-5d048f4ec520\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.807694 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-combined-ca-bundle\") pod \"022b0b02-d8bb-4560-af50-5d048f4ec520\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.807759 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-ssh-key\") pod \"022b0b02-d8bb-4560-af50-5d048f4ec520\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.808468 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-secret-0\") pod \"022b0b02-d8bb-4560-af50-5d048f4ec520\" (UID: \"022b0b02-d8bb-4560-af50-5d048f4ec520\") " Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.814371 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/022b0b02-d8bb-4560-af50-5d048f4ec520-kube-api-access-8dvwf" (OuterVolumeSpecName: "kube-api-access-8dvwf") pod "022b0b02-d8bb-4560-af50-5d048f4ec520" (UID: "022b0b02-d8bb-4560-af50-5d048f4ec520"). InnerVolumeSpecName "kube-api-access-8dvwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.815852 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "022b0b02-d8bb-4560-af50-5d048f4ec520" (UID: "022b0b02-d8bb-4560-af50-5d048f4ec520"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.856776 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-inventory" (OuterVolumeSpecName: "inventory") pod "022b0b02-d8bb-4560-af50-5d048f4ec520" (UID: "022b0b02-d8bb-4560-af50-5d048f4ec520"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.859892 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "022b0b02-d8bb-4560-af50-5d048f4ec520" (UID: "022b0b02-d8bb-4560-af50-5d048f4ec520"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.864038 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "022b0b02-d8bb-4560-af50-5d048f4ec520" (UID: "022b0b02-d8bb-4560-af50-5d048f4ec520"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.911559 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.911804 4573 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.911894 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.912064 4573 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/022b0b02-d8bb-4560-af50-5d048f4ec520-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:24:53 crc kubenswrapper[4573]: I1203 09:24:53.912142 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dvwf\" (UniqueName: \"kubernetes.io/projected/022b0b02-d8bb-4560-af50-5d048f4ec520-kube-api-access-8dvwf\") on node \"crc\" DevicePath \"\"" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.325565 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" event={"ID":"022b0b02-d8bb-4560-af50-5d048f4ec520","Type":"ContainerDied","Data":"9a2875012ff3ba7a96456da2c61ff54f954c808499c625db82aba7c4dfed9b58"} Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.325646 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a2875012ff3ba7a96456da2c61ff54f954c808499c625db82aba7c4dfed9b58" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.325651 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.521528 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt"] Dec 03 09:24:54 crc kubenswrapper[4573]: E1203 09:24:54.522406 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerName="registry-server" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522425 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerName="registry-server" Dec 03 09:24:54 crc kubenswrapper[4573]: E1203 09:24:54.522438 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="022b0b02-d8bb-4560-af50-5d048f4ec520" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522466 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="022b0b02-d8bb-4560-af50-5d048f4ec520" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 09:24:54 crc kubenswrapper[4573]: E1203 09:24:54.522488 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerName="registry-server" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522494 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerName="registry-server" Dec 03 09:24:54 crc kubenswrapper[4573]: E1203 09:24:54.522516 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerName="extract-content" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522522 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerName="extract-content" Dec 03 09:24:54 crc kubenswrapper[4573]: E1203 09:24:54.522558 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerName="extract-utilities" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522565 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerName="extract-utilities" Dec 03 09:24:54 crc kubenswrapper[4573]: E1203 09:24:54.522579 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerName="extract-utilities" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522585 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerName="extract-utilities" Dec 03 09:24:54 crc kubenswrapper[4573]: E1203 09:24:54.522597 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerName="extract-content" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522602 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerName="extract-content" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522879 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="85bba5be-441a-4f77-9a01-5a191e13ab6c" containerName="registry-server" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522929 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1558ad3-c1c5-4760-81f4-40040b79096f" containerName="registry-server" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.522946 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="022b0b02-d8bb-4560-af50-5d048f4ec520" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.523707 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.527315 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.527487 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.527626 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.531130 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.531454 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.531497 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.538225 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.540856 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt"] Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.555446 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.555692 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.555820 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.555955 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.556037 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.556169 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4pwr\" (UniqueName: \"kubernetes.io/projected/5847f67b-3130-4fa9-831f-bcfe0c27ad48-kube-api-access-p4pwr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.556238 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.556355 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.556449 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658045 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658143 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658167 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658212 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4pwr\" (UniqueName: \"kubernetes.io/projected/5847f67b-3130-4fa9-831f-bcfe0c27ad48-kube-api-access-p4pwr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658253 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658296 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658352 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658376 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.658414 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.660343 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.665399 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.665926 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.666440 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.667875 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.670013 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.673800 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.695281 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.702416 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4pwr\" (UniqueName: \"kubernetes.io/projected/5847f67b-3130-4fa9-831f-bcfe0c27ad48-kube-api-access-p4pwr\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jc6rt\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:54 crc kubenswrapper[4573]: I1203 09:24:54.845550 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:24:55 crc kubenswrapper[4573]: I1203 09:24:55.457220 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt"] Dec 03 09:24:56 crc kubenswrapper[4573]: I1203 09:24:56.349662 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" event={"ID":"5847f67b-3130-4fa9-831f-bcfe0c27ad48","Type":"ContainerStarted","Data":"c6a3773f8a610eb30b4ba248e64a93c824d7dbb986bd0e6e188bcfc592f68d82"} Dec 03 09:24:56 crc kubenswrapper[4573]: I1203 09:24:56.350065 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" event={"ID":"5847f67b-3130-4fa9-831f-bcfe0c27ad48","Type":"ContainerStarted","Data":"1511eb866ba3a256b51ddaf1ef9196fa651bd986089e8799c347a0ed238b1fd2"} Dec 03 09:24:56 crc kubenswrapper[4573]: I1203 09:24:56.375005 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" podStartSLOduration=2.197019012 podStartE2EDuration="2.374941413s" podCreationTimestamp="2025-12-03 09:24:54 +0000 UTC" firstStartedPulling="2025-12-03 09:24:55.471688606 +0000 UTC m=+2816.040067865" lastFinishedPulling="2025-12-03 09:24:55.649611007 +0000 UTC m=+2816.217990266" observedRunningTime="2025-12-03 09:24:56.372697732 +0000 UTC m=+2816.941076991" watchObservedRunningTime="2025-12-03 09:24:56.374941413 +0000 UTC m=+2816.943320672" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.083177 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cjggp"] Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.087395 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.111649 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjggp"] Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.136602 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-utilities\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.136667 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-catalog-content\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.136708 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq2jv\" (UniqueName: \"kubernetes.io/projected/c5ee9202-6f7f-4534-adef-bbc2124b2a56-kube-api-access-nq2jv\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.238903 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-utilities\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.238964 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-catalog-content\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.239001 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq2jv\" (UniqueName: \"kubernetes.io/projected/c5ee9202-6f7f-4534-adef-bbc2124b2a56-kube-api-access-nq2jv\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.239915 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-utilities\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.240176 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-catalog-content\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.265142 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq2jv\" (UniqueName: \"kubernetes.io/projected/c5ee9202-6f7f-4534-adef-bbc2124b2a56-kube-api-access-nq2jv\") pod \"certified-operators-cjggp\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:02 crc kubenswrapper[4573]: I1203 09:25:02.418613 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:03 crc kubenswrapper[4573]: I1203 09:25:03.061543 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjggp"] Dec 03 09:25:03 crc kubenswrapper[4573]: I1203 09:25:03.424667 4573 generic.go:334] "Generic (PLEG): container finished" podID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerID="c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1" exitCode=0 Dec 03 09:25:03 crc kubenswrapper[4573]: I1203 09:25:03.424764 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjggp" event={"ID":"c5ee9202-6f7f-4534-adef-bbc2124b2a56","Type":"ContainerDied","Data":"c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1"} Dec 03 09:25:03 crc kubenswrapper[4573]: I1203 09:25:03.426356 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjggp" event={"ID":"c5ee9202-6f7f-4534-adef-bbc2124b2a56","Type":"ContainerStarted","Data":"c063d3ee863a29ebaa1947d8aa38df7d94f246a610c86f99dcb2102f7ab42d67"} Dec 03 09:25:04 crc kubenswrapper[4573]: I1203 09:25:04.442493 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjggp" event={"ID":"c5ee9202-6f7f-4534-adef-bbc2124b2a56","Type":"ContainerStarted","Data":"01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6"} Dec 03 09:25:06 crc kubenswrapper[4573]: I1203 09:25:06.480494 4573 generic.go:334] "Generic (PLEG): container finished" podID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerID="01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6" exitCode=0 Dec 03 09:25:06 crc kubenswrapper[4573]: I1203 09:25:06.480568 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjggp" event={"ID":"c5ee9202-6f7f-4534-adef-bbc2124b2a56","Type":"ContainerDied","Data":"01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6"} Dec 03 09:25:08 crc kubenswrapper[4573]: I1203 09:25:08.518831 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjggp" event={"ID":"c5ee9202-6f7f-4534-adef-bbc2124b2a56","Type":"ContainerStarted","Data":"39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61"} Dec 03 09:25:08 crc kubenswrapper[4573]: I1203 09:25:08.564785 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cjggp" podStartSLOduration=2.3870670609999998 podStartE2EDuration="6.564753343s" podCreationTimestamp="2025-12-03 09:25:02 +0000 UTC" firstStartedPulling="2025-12-03 09:25:03.427517108 +0000 UTC m=+2823.995896367" lastFinishedPulling="2025-12-03 09:25:07.60520339 +0000 UTC m=+2828.173582649" observedRunningTime="2025-12-03 09:25:08.553106045 +0000 UTC m=+2829.121485314" watchObservedRunningTime="2025-12-03 09:25:08.564753343 +0000 UTC m=+2829.133132612" Dec 03 09:25:12 crc kubenswrapper[4573]: I1203 09:25:12.418746 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:12 crc kubenswrapper[4573]: I1203 09:25:12.419312 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:12 crc kubenswrapper[4573]: I1203 09:25:12.489862 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:12 crc kubenswrapper[4573]: I1203 09:25:12.611377 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:12 crc kubenswrapper[4573]: I1203 09:25:12.733175 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjggp"] Dec 03 09:25:14 crc kubenswrapper[4573]: I1203 09:25:14.587605 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cjggp" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerName="registry-server" containerID="cri-o://39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61" gracePeriod=2 Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.092948 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.095834 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-catalog-content\") pod \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.095888 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-utilities\") pod \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.096111 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq2jv\" (UniqueName: \"kubernetes.io/projected/c5ee9202-6f7f-4534-adef-bbc2124b2a56-kube-api-access-nq2jv\") pod \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\" (UID: \"c5ee9202-6f7f-4534-adef-bbc2124b2a56\") " Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.096733 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-utilities" (OuterVolumeSpecName: "utilities") pod "c5ee9202-6f7f-4534-adef-bbc2124b2a56" (UID: "c5ee9202-6f7f-4534-adef-bbc2124b2a56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.097750 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.104001 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ee9202-6f7f-4534-adef-bbc2124b2a56-kube-api-access-nq2jv" (OuterVolumeSpecName: "kube-api-access-nq2jv") pod "c5ee9202-6f7f-4534-adef-bbc2124b2a56" (UID: "c5ee9202-6f7f-4534-adef-bbc2124b2a56"). InnerVolumeSpecName "kube-api-access-nq2jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.182272 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5ee9202-6f7f-4534-adef-bbc2124b2a56" (UID: "c5ee9202-6f7f-4534-adef-bbc2124b2a56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.198715 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq2jv\" (UniqueName: \"kubernetes.io/projected/c5ee9202-6f7f-4534-adef-bbc2124b2a56-kube-api-access-nq2jv\") on node \"crc\" DevicePath \"\"" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.198777 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ee9202-6f7f-4534-adef-bbc2124b2a56-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.601600 4573 generic.go:334] "Generic (PLEG): container finished" podID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerID="39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61" exitCode=0 Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.601680 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjggp" event={"ID":"c5ee9202-6f7f-4534-adef-bbc2124b2a56","Type":"ContainerDied","Data":"39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61"} Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.601728 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjggp" event={"ID":"c5ee9202-6f7f-4534-adef-bbc2124b2a56","Type":"ContainerDied","Data":"c063d3ee863a29ebaa1947d8aa38df7d94f246a610c86f99dcb2102f7ab42d67"} Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.601835 4573 scope.go:117] "RemoveContainer" containerID="39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.602142 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjggp" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.631282 4573 scope.go:117] "RemoveContainer" containerID="01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.655147 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjggp"] Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.666721 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cjggp"] Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.732294 4573 scope.go:117] "RemoveContainer" containerID="c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.763809 4573 scope.go:117] "RemoveContainer" containerID="39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61" Dec 03 09:25:15 crc kubenswrapper[4573]: E1203 09:25:15.764374 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61\": container with ID starting with 39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61 not found: ID does not exist" containerID="39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.764414 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61"} err="failed to get container status \"39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61\": rpc error: code = NotFound desc = could not find container \"39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61\": container with ID starting with 39eb957b63baa3e61e4dc8f437f4b54d7fd705f9a5cf492a8f73b5c2b9faad61 not found: ID does not exist" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.764446 4573 scope.go:117] "RemoveContainer" containerID="01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6" Dec 03 09:25:15 crc kubenswrapper[4573]: E1203 09:25:15.764831 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6\": container with ID starting with 01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6 not found: ID does not exist" containerID="01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.764946 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6"} err="failed to get container status \"01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6\": rpc error: code = NotFound desc = could not find container \"01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6\": container with ID starting with 01281598ec7a0a5a18ba8eeba0ed83c0e47b20573c65f03e0d7f5e66a35e29f6 not found: ID does not exist" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.765076 4573 scope.go:117] "RemoveContainer" containerID="c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1" Dec 03 09:25:15 crc kubenswrapper[4573]: E1203 09:25:15.765452 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1\": container with ID starting with c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1 not found: ID does not exist" containerID="c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1" Dec 03 09:25:15 crc kubenswrapper[4573]: I1203 09:25:15.765473 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1"} err="failed to get container status \"c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1\": rpc error: code = NotFound desc = could not find container \"c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1\": container with ID starting with c2c15f6f6c906d48dc6ff22cc81df792d11ff66c3912acc015e7e4e8bb51b9f1 not found: ID does not exist" Dec 03 09:25:16 crc kubenswrapper[4573]: I1203 09:25:16.043900 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" path="/var/lib/kubelet/pods/c5ee9202-6f7f-4534-adef-bbc2124b2a56/volumes" Dec 03 09:25:26 crc kubenswrapper[4573]: I1203 09:25:26.943691 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:25:26 crc kubenswrapper[4573]: I1203 09:25:26.944481 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:25:56 crc kubenswrapper[4573]: I1203 09:25:56.943236 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:25:56 crc kubenswrapper[4573]: I1203 09:25:56.944092 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:26:26 crc kubenswrapper[4573]: I1203 09:26:26.943343 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:26:26 crc kubenswrapper[4573]: I1203 09:26:26.943868 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:26:26 crc kubenswrapper[4573]: I1203 09:26:26.943912 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:26:26 crc kubenswrapper[4573]: I1203 09:26:26.944757 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9661b8f523a73ea76d82ecaf1583ed165c00467787ee991495f624ee4ae4fbd1"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:26:26 crc kubenswrapper[4573]: I1203 09:26:26.944828 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://9661b8f523a73ea76d82ecaf1583ed165c00467787ee991495f624ee4ae4fbd1" gracePeriod=600 Dec 03 09:26:27 crc kubenswrapper[4573]: I1203 09:26:27.402430 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="9661b8f523a73ea76d82ecaf1583ed165c00467787ee991495f624ee4ae4fbd1" exitCode=0 Dec 03 09:26:27 crc kubenswrapper[4573]: I1203 09:26:27.402518 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"9661b8f523a73ea76d82ecaf1583ed165c00467787ee991495f624ee4ae4fbd1"} Dec 03 09:26:27 crc kubenswrapper[4573]: I1203 09:26:27.402686 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8"} Dec 03 09:26:27 crc kubenswrapper[4573]: I1203 09:26:27.402706 4573 scope.go:117] "RemoveContainer" containerID="bec8f1ce8231c411d1d8a1e2623dbf5b0f8ecb9b91109404a2fc88325c4ccba3" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.345436 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qfxft"] Dec 03 09:27:06 crc kubenswrapper[4573]: E1203 09:27:06.346789 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerName="extract-utilities" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.346814 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerName="extract-utilities" Dec 03 09:27:06 crc kubenswrapper[4573]: E1203 09:27:06.346857 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerName="registry-server" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.346869 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerName="registry-server" Dec 03 09:27:06 crc kubenswrapper[4573]: E1203 09:27:06.346910 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerName="extract-content" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.346921 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerName="extract-content" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.347267 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ee9202-6f7f-4534-adef-bbc2124b2a56" containerName="registry-server" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.349294 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.373415 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qfxft"] Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.417153 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-catalog-content\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.417313 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-utilities\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.417387 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54w4z\" (UniqueName: \"kubernetes.io/projected/3bee78ab-73c6-4930-9db5-a8059953605f-kube-api-access-54w4z\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.519463 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-catalog-content\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.519584 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-utilities\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.519622 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54w4z\" (UniqueName: \"kubernetes.io/projected/3bee78ab-73c6-4930-9db5-a8059953605f-kube-api-access-54w4z\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.520104 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-catalog-content\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.520150 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-utilities\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.543183 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54w4z\" (UniqueName: \"kubernetes.io/projected/3bee78ab-73c6-4930-9db5-a8059953605f-kube-api-access-54w4z\") pod \"redhat-operators-qfxft\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:06 crc kubenswrapper[4573]: I1203 09:27:06.695994 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:07 crc kubenswrapper[4573]: I1203 09:27:07.231824 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qfxft"] Dec 03 09:27:07 crc kubenswrapper[4573]: I1203 09:27:07.862581 4573 generic.go:334] "Generic (PLEG): container finished" podID="3bee78ab-73c6-4930-9db5-a8059953605f" containerID="0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018" exitCode=0 Dec 03 09:27:07 crc kubenswrapper[4573]: I1203 09:27:07.862784 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfxft" event={"ID":"3bee78ab-73c6-4930-9db5-a8059953605f","Type":"ContainerDied","Data":"0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018"} Dec 03 09:27:07 crc kubenswrapper[4573]: I1203 09:27:07.864074 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfxft" event={"ID":"3bee78ab-73c6-4930-9db5-a8059953605f","Type":"ContainerStarted","Data":"7f92d4f12341d25f00606dde864fba267c54865d4dab698f409866c178d222cb"} Dec 03 09:27:07 crc kubenswrapper[4573]: I1203 09:27:07.865665 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:27:09 crc kubenswrapper[4573]: I1203 09:27:09.898679 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfxft" event={"ID":"3bee78ab-73c6-4930-9db5-a8059953605f","Type":"ContainerStarted","Data":"bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec"} Dec 03 09:27:14 crc kubenswrapper[4573]: I1203 09:27:14.953804 4573 generic.go:334] "Generic (PLEG): container finished" podID="3bee78ab-73c6-4930-9db5-a8059953605f" containerID="bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec" exitCode=0 Dec 03 09:27:14 crc kubenswrapper[4573]: I1203 09:27:14.953915 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfxft" event={"ID":"3bee78ab-73c6-4930-9db5-a8059953605f","Type":"ContainerDied","Data":"bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec"} Dec 03 09:27:16 crc kubenswrapper[4573]: I1203 09:27:16.976460 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfxft" event={"ID":"3bee78ab-73c6-4930-9db5-a8059953605f","Type":"ContainerStarted","Data":"003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544"} Dec 03 09:27:16 crc kubenswrapper[4573]: I1203 09:27:16.998537 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qfxft" podStartSLOduration=2.5146758460000003 podStartE2EDuration="10.998514986s" podCreationTimestamp="2025-12-03 09:27:06 +0000 UTC" firstStartedPulling="2025-12-03 09:27:07.865329473 +0000 UTC m=+2948.433708732" lastFinishedPulling="2025-12-03 09:27:16.349168613 +0000 UTC m=+2956.917547872" observedRunningTime="2025-12-03 09:27:16.99719179 +0000 UTC m=+2957.565571099" watchObservedRunningTime="2025-12-03 09:27:16.998514986 +0000 UTC m=+2957.566894265" Dec 03 09:27:26 crc kubenswrapper[4573]: I1203 09:27:26.699360 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:26 crc kubenswrapper[4573]: I1203 09:27:26.700171 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:27 crc kubenswrapper[4573]: I1203 09:27:27.760482 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qfxft" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="registry-server" probeResult="failure" output=< Dec 03 09:27:27 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:27:27 crc kubenswrapper[4573]: > Dec 03 09:27:36 crc kubenswrapper[4573]: I1203 09:27:36.779137 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:36 crc kubenswrapper[4573]: I1203 09:27:36.853063 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:37 crc kubenswrapper[4573]: I1203 09:27:37.546009 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qfxft"] Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.191085 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qfxft" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="registry-server" containerID="cri-o://003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544" gracePeriod=2 Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.739836 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.883794 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54w4z\" (UniqueName: \"kubernetes.io/projected/3bee78ab-73c6-4930-9db5-a8059953605f-kube-api-access-54w4z\") pod \"3bee78ab-73c6-4930-9db5-a8059953605f\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.884870 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-utilities\") pod \"3bee78ab-73c6-4930-9db5-a8059953605f\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.884903 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-catalog-content\") pod \"3bee78ab-73c6-4930-9db5-a8059953605f\" (UID: \"3bee78ab-73c6-4930-9db5-a8059953605f\") " Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.885537 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-utilities" (OuterVolumeSpecName: "utilities") pod "3bee78ab-73c6-4930-9db5-a8059953605f" (UID: "3bee78ab-73c6-4930-9db5-a8059953605f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.890521 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bee78ab-73c6-4930-9db5-a8059953605f-kube-api-access-54w4z" (OuterVolumeSpecName: "kube-api-access-54w4z") pod "3bee78ab-73c6-4930-9db5-a8059953605f" (UID: "3bee78ab-73c6-4930-9db5-a8059953605f"). InnerVolumeSpecName "kube-api-access-54w4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.987364 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54w4z\" (UniqueName: \"kubernetes.io/projected/3bee78ab-73c6-4930-9db5-a8059953605f-kube-api-access-54w4z\") on node \"crc\" DevicePath \"\"" Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.987402 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:27:38 crc kubenswrapper[4573]: I1203 09:27:38.998408 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bee78ab-73c6-4930-9db5-a8059953605f" (UID: "3bee78ab-73c6-4930-9db5-a8059953605f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.091099 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bee78ab-73c6-4930-9db5-a8059953605f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.205797 4573 generic.go:334] "Generic (PLEG): container finished" podID="3bee78ab-73c6-4930-9db5-a8059953605f" containerID="003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544" exitCode=0 Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.205843 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfxft" event={"ID":"3bee78ab-73c6-4930-9db5-a8059953605f","Type":"ContainerDied","Data":"003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544"} Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.205868 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qfxft" event={"ID":"3bee78ab-73c6-4930-9db5-a8059953605f","Type":"ContainerDied","Data":"7f92d4f12341d25f00606dde864fba267c54865d4dab698f409866c178d222cb"} Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.205885 4573 scope.go:117] "RemoveContainer" containerID="003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.206033 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qfxft" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.246471 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qfxft"] Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.255636 4573 scope.go:117] "RemoveContainer" containerID="bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.262359 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qfxft"] Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.291007 4573 scope.go:117] "RemoveContainer" containerID="0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.396219 4573 scope.go:117] "RemoveContainer" containerID="003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544" Dec 03 09:27:39 crc kubenswrapper[4573]: E1203 09:27:39.396753 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544\": container with ID starting with 003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544 not found: ID does not exist" containerID="003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.396817 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544"} err="failed to get container status \"003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544\": rpc error: code = NotFound desc = could not find container \"003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544\": container with ID starting with 003f92bfa87fb6ea6e443f2baa20f18c93a84554593f30628ceff398c5a0d544 not found: ID does not exist" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.396867 4573 scope.go:117] "RemoveContainer" containerID="bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec" Dec 03 09:27:39 crc kubenswrapper[4573]: E1203 09:27:39.397308 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec\": container with ID starting with bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec not found: ID does not exist" containerID="bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.397344 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec"} err="failed to get container status \"bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec\": rpc error: code = NotFound desc = could not find container \"bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec\": container with ID starting with bc5c6eacba28e1283ccf6fd36f5f2707fac12f5be2f595c9a5d05e0cda9867ec not found: ID does not exist" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.397382 4573 scope.go:117] "RemoveContainer" containerID="0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018" Dec 03 09:27:39 crc kubenswrapper[4573]: E1203 09:27:39.397711 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018\": container with ID starting with 0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018 not found: ID does not exist" containerID="0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018" Dec 03 09:27:39 crc kubenswrapper[4573]: I1203 09:27:39.397777 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018"} err="failed to get container status \"0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018\": rpc error: code = NotFound desc = could not find container \"0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018\": container with ID starting with 0854f2e8dae6a1b4da7b19dc884b1ee797ed38bda4db7903f7b6745927f96018 not found: ID does not exist" Dec 03 09:27:40 crc kubenswrapper[4573]: I1203 09:27:40.057174 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" path="/var/lib/kubelet/pods/3bee78ab-73c6-4930-9db5-a8059953605f/volumes" Dec 03 09:28:09 crc kubenswrapper[4573]: I1203 09:28:09.569115 4573 generic.go:334] "Generic (PLEG): container finished" podID="5847f67b-3130-4fa9-831f-bcfe0c27ad48" containerID="c6a3773f8a610eb30b4ba248e64a93c824d7dbb986bd0e6e188bcfc592f68d82" exitCode=0 Dec 03 09:28:09 crc kubenswrapper[4573]: I1203 09:28:09.569255 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" event={"ID":"5847f67b-3130-4fa9-831f-bcfe0c27ad48","Type":"ContainerDied","Data":"c6a3773f8a610eb30b4ba248e64a93c824d7dbb986bd0e6e188bcfc592f68d82"} Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.001373 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.146606 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-1\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.146768 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-0\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.146840 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4pwr\" (UniqueName: \"kubernetes.io/projected/5847f67b-3130-4fa9-831f-bcfe0c27ad48-kube-api-access-p4pwr\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.146895 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-1\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.146968 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-ssh-key\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.147094 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-0\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.147164 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-inventory\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.147928 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-combined-ca-bundle\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.148000 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-extra-config-0\") pod \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\" (UID: \"5847f67b-3130-4fa9-831f-bcfe0c27ad48\") " Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.154298 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5847f67b-3130-4fa9-831f-bcfe0c27ad48-kube-api-access-p4pwr" (OuterVolumeSpecName: "kube-api-access-p4pwr") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "kube-api-access-p4pwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.155181 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.180627 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.210364 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.210878 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.211892 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.212849 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.214665 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-inventory" (OuterVolumeSpecName: "inventory") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.220584 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "5847f67b-3130-4fa9-831f-bcfe0c27ad48" (UID: "5847f67b-3130-4fa9-831f-bcfe0c27ad48"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251443 4573 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251579 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4pwr\" (UniqueName: \"kubernetes.io/projected/5847f67b-3130-4fa9-831f-bcfe0c27ad48-kube-api-access-p4pwr\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251594 4573 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251607 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251620 4573 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251632 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251648 4573 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251661 4573 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.251672 4573 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5847f67b-3130-4fa9-831f-bcfe0c27ad48-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.588977 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" event={"ID":"5847f67b-3130-4fa9-831f-bcfe0c27ad48","Type":"ContainerDied","Data":"1511eb866ba3a256b51ddaf1ef9196fa651bd986089e8799c347a0ed238b1fd2"} Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.589033 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1511eb866ba3a256b51ddaf1ef9196fa651bd986089e8799c347a0ed238b1fd2" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.589116 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jc6rt" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.779710 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc"] Dec 03 09:28:11 crc kubenswrapper[4573]: E1203 09:28:11.780591 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5847f67b-3130-4fa9-831f-bcfe0c27ad48" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.780613 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="5847f67b-3130-4fa9-831f-bcfe0c27ad48" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 09:28:11 crc kubenswrapper[4573]: E1203 09:28:11.780646 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="extract-content" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.780654 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="extract-content" Dec 03 09:28:11 crc kubenswrapper[4573]: E1203 09:28:11.780672 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="registry-server" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.780679 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="registry-server" Dec 03 09:28:11 crc kubenswrapper[4573]: E1203 09:28:11.780694 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="extract-utilities" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.780700 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="extract-utilities" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.780904 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="5847f67b-3130-4fa9-831f-bcfe0c27ad48" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.780919 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bee78ab-73c6-4930-9db5-a8059953605f" containerName="registry-server" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.781809 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.786337 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.786671 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.788492 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.789370 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.798084 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-zm6tz" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.800628 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc"] Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.863907 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.863986 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.864023 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.864061 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.864101 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.864147 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.864218 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmjzm\" (UniqueName: \"kubernetes.io/projected/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-kube-api-access-pmjzm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.966275 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.966441 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmjzm\" (UniqueName: \"kubernetes.io/projected/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-kube-api-access-pmjzm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.966521 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.966592 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.966684 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.966737 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.966783 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.973751 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.974815 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.975351 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.977994 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.978677 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.984724 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:11 crc kubenswrapper[4573]: I1203 09:28:11.991994 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmjzm\" (UniqueName: \"kubernetes.io/projected/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-kube-api-access-pmjzm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:12 crc kubenswrapper[4573]: I1203 09:28:12.105573 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:28:12 crc kubenswrapper[4573]: I1203 09:28:12.702417 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc"] Dec 03 09:28:13 crc kubenswrapper[4573]: I1203 09:28:13.616191 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" event={"ID":"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9","Type":"ContainerStarted","Data":"159caac13e26af8ff7e20d53114466a89549bff12419c0bf4713ed55172141e9"} Dec 03 09:28:17 crc kubenswrapper[4573]: I1203 09:28:17.673111 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" event={"ID":"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9","Type":"ContainerStarted","Data":"2d5748f4e30bf052b251c696f99372421a8ba169f5e7f1ab9b1c70d1a7369f78"} Dec 03 09:28:18 crc kubenswrapper[4573]: I1203 09:28:18.714880 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" podStartSLOduration=3.513305969 podStartE2EDuration="7.714856803s" podCreationTimestamp="2025-12-03 09:28:11 +0000 UTC" firstStartedPulling="2025-12-03 09:28:12.715973396 +0000 UTC m=+3013.284352655" lastFinishedPulling="2025-12-03 09:28:16.91752422 +0000 UTC m=+3017.485903489" observedRunningTime="2025-12-03 09:28:18.705146259 +0000 UTC m=+3019.273525538" watchObservedRunningTime="2025-12-03 09:28:18.714856803 +0000 UTC m=+3019.283236062" Dec 03 09:28:56 crc kubenswrapper[4573]: I1203 09:28:56.943705 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:28:56 crc kubenswrapper[4573]: I1203 09:28:56.944419 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:29:26 crc kubenswrapper[4573]: I1203 09:29:26.942935 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:29:26 crc kubenswrapper[4573]: I1203 09:29:26.943709 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:29:56 crc kubenswrapper[4573]: I1203 09:29:56.943855 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:29:56 crc kubenswrapper[4573]: I1203 09:29:56.944786 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:29:56 crc kubenswrapper[4573]: I1203 09:29:56.944871 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:29:56 crc kubenswrapper[4573]: I1203 09:29:56.946179 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:29:56 crc kubenswrapper[4573]: I1203 09:29:56.946293 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" gracePeriod=600 Dec 03 09:29:57 crc kubenswrapper[4573]: E1203 09:29:57.079889 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:29:57 crc kubenswrapper[4573]: I1203 09:29:57.731986 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" exitCode=0 Dec 03 09:29:57 crc kubenswrapper[4573]: I1203 09:29:57.732280 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8"} Dec 03 09:29:57 crc kubenswrapper[4573]: I1203 09:29:57.732464 4573 scope.go:117] "RemoveContainer" containerID="9661b8f523a73ea76d82ecaf1583ed165c00467787ee991495f624ee4ae4fbd1" Dec 03 09:29:57 crc kubenswrapper[4573]: I1203 09:29:57.733379 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:29:57 crc kubenswrapper[4573]: E1203 09:29:57.733708 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.172089 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn"] Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.174877 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.179106 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.183401 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.185991 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn"] Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.325525 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46e2a8e9-4614-4763-979a-daa41262f708-secret-volume\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.325584 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp4w2\" (UniqueName: \"kubernetes.io/projected/46e2a8e9-4614-4763-979a-daa41262f708-kube-api-access-hp4w2\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.325732 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46e2a8e9-4614-4763-979a-daa41262f708-config-volume\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.428085 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46e2a8e9-4614-4763-979a-daa41262f708-secret-volume\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.428150 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp4w2\" (UniqueName: \"kubernetes.io/projected/46e2a8e9-4614-4763-979a-daa41262f708-kube-api-access-hp4w2\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.428263 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46e2a8e9-4614-4763-979a-daa41262f708-config-volume\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.429434 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46e2a8e9-4614-4763-979a-daa41262f708-config-volume\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.437708 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46e2a8e9-4614-4763-979a-daa41262f708-secret-volume\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.454799 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp4w2\" (UniqueName: \"kubernetes.io/projected/46e2a8e9-4614-4763-979a-daa41262f708-kube-api-access-hp4w2\") pod \"collect-profiles-29412570-q7ssn\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.501083 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:00 crc kubenswrapper[4573]: I1203 09:30:00.971616 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn"] Dec 03 09:30:01 crc kubenswrapper[4573]: I1203 09:30:01.792119 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" event={"ID":"46e2a8e9-4614-4763-979a-daa41262f708","Type":"ContainerStarted","Data":"46bf909f92d1673520987d5085f117ec07075c40174461ec82cc9a2cb2b16bbb"} Dec 03 09:30:02 crc kubenswrapper[4573]: I1203 09:30:02.812103 4573 generic.go:334] "Generic (PLEG): container finished" podID="46e2a8e9-4614-4763-979a-daa41262f708" containerID="29212cc6ebedf6eb1bdc4823a7236850851d7da785826f168634e249af29ed1b" exitCode=0 Dec 03 09:30:02 crc kubenswrapper[4573]: I1203 09:30:02.812611 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" event={"ID":"46e2a8e9-4614-4763-979a-daa41262f708","Type":"ContainerDied","Data":"29212cc6ebedf6eb1bdc4823a7236850851d7da785826f168634e249af29ed1b"} Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.192606 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.326040 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46e2a8e9-4614-4763-979a-daa41262f708-config-volume\") pod \"46e2a8e9-4614-4763-979a-daa41262f708\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.326117 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp4w2\" (UniqueName: \"kubernetes.io/projected/46e2a8e9-4614-4763-979a-daa41262f708-kube-api-access-hp4w2\") pod \"46e2a8e9-4614-4763-979a-daa41262f708\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.326159 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46e2a8e9-4614-4763-979a-daa41262f708-secret-volume\") pod \"46e2a8e9-4614-4763-979a-daa41262f708\" (UID: \"46e2a8e9-4614-4763-979a-daa41262f708\") " Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.326862 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46e2a8e9-4614-4763-979a-daa41262f708-config-volume" (OuterVolumeSpecName: "config-volume") pod "46e2a8e9-4614-4763-979a-daa41262f708" (UID: "46e2a8e9-4614-4763-979a-daa41262f708"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.327644 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46e2a8e9-4614-4763-979a-daa41262f708-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.333177 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46e2a8e9-4614-4763-979a-daa41262f708-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "46e2a8e9-4614-4763-979a-daa41262f708" (UID: "46e2a8e9-4614-4763-979a-daa41262f708"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.334930 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46e2a8e9-4614-4763-979a-daa41262f708-kube-api-access-hp4w2" (OuterVolumeSpecName: "kube-api-access-hp4w2") pod "46e2a8e9-4614-4763-979a-daa41262f708" (UID: "46e2a8e9-4614-4763-979a-daa41262f708"). InnerVolumeSpecName "kube-api-access-hp4w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.429055 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp4w2\" (UniqueName: \"kubernetes.io/projected/46e2a8e9-4614-4763-979a-daa41262f708-kube-api-access-hp4w2\") on node \"crc\" DevicePath \"\"" Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.429105 4573 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46e2a8e9-4614-4763-979a-daa41262f708-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.835445 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" event={"ID":"46e2a8e9-4614-4763-979a-daa41262f708","Type":"ContainerDied","Data":"46bf909f92d1673520987d5085f117ec07075c40174461ec82cc9a2cb2b16bbb"} Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.835502 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46bf909f92d1673520987d5085f117ec07075c40174461ec82cc9a2cb2b16bbb" Dec 03 09:30:04 crc kubenswrapper[4573]: I1203 09:30:04.835572 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn" Dec 03 09:30:05 crc kubenswrapper[4573]: I1203 09:30:05.290824 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph"] Dec 03 09:30:05 crc kubenswrapper[4573]: I1203 09:30:05.301469 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412525-rvqph"] Dec 03 09:30:06 crc kubenswrapper[4573]: I1203 09:30:06.049831 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc501733-940b-4735-8607-31f0bc1b1297" path="/var/lib/kubelet/pods/cc501733-940b-4735-8607-31f0bc1b1297/volumes" Dec 03 09:30:08 crc kubenswrapper[4573]: I1203 09:30:08.430149 4573 scope.go:117] "RemoveContainer" containerID="3be2fafc5e13740d824be85dada8f6f0568fbfa608c67f34f15419d1a5bc037a" Dec 03 09:30:13 crc kubenswrapper[4573]: I1203 09:30:13.031425 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:30:13 crc kubenswrapper[4573]: E1203 09:30:13.032784 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:30:26 crc kubenswrapper[4573]: I1203 09:30:26.030799 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:30:26 crc kubenswrapper[4573]: E1203 09:30:26.032095 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:30:40 crc kubenswrapper[4573]: I1203 09:30:40.037220 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:30:40 crc kubenswrapper[4573]: E1203 09:30:40.038080 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:30:53 crc kubenswrapper[4573]: I1203 09:30:53.032112 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:30:53 crc kubenswrapper[4573]: E1203 09:30:53.033524 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:31:08 crc kubenswrapper[4573]: I1203 09:31:08.030170 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:31:08 crc kubenswrapper[4573]: E1203 09:31:08.031055 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:31:22 crc kubenswrapper[4573]: I1203 09:31:22.031236 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:31:22 crc kubenswrapper[4573]: E1203 09:31:22.032415 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:31:34 crc kubenswrapper[4573]: I1203 09:31:34.031650 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:31:34 crc kubenswrapper[4573]: E1203 09:31:34.035135 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:31:48 crc kubenswrapper[4573]: I1203 09:31:48.031067 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:31:48 crc kubenswrapper[4573]: E1203 09:31:48.032464 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:31:51 crc kubenswrapper[4573]: I1203 09:31:51.946076 4573 generic.go:334] "Generic (PLEG): container finished" podID="fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" containerID="2d5748f4e30bf052b251c696f99372421a8ba169f5e7f1ab9b1c70d1a7369f78" exitCode=0 Dec 03 09:31:51 crc kubenswrapper[4573]: I1203 09:31:51.946873 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" event={"ID":"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9","Type":"ContainerDied","Data":"2d5748f4e30bf052b251c696f99372421a8ba169f5e7f1ab9b1c70d1a7369f78"} Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.488018 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.626612 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmjzm\" (UniqueName: \"kubernetes.io/projected/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-kube-api-access-pmjzm\") pod \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.627137 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ssh-key\") pod \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.627170 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-2\") pod \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.627194 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-inventory\") pod \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.627217 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-1\") pod \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.627242 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-0\") pod \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.627285 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-telemetry-combined-ca-bundle\") pod \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\" (UID: \"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9\") " Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.634559 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" (UID: "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.635390 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-kube-api-access-pmjzm" (OuterVolumeSpecName: "kube-api-access-pmjzm") pod "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" (UID: "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9"). InnerVolumeSpecName "kube-api-access-pmjzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.655472 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" (UID: "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.660587 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" (UID: "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.672694 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-inventory" (OuterVolumeSpecName: "inventory") pod "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" (UID: "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.699116 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" (UID: "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.714082 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" (UID: "fa5259c8-c545-4b45-a9c6-e0bca4d3aba9"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.729275 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.729554 4573 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.729658 4573 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.729765 4573 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.729854 4573 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.729938 4573 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.730022 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmjzm\" (UniqueName: \"kubernetes.io/projected/fa5259c8-c545-4b45-a9c6-e0bca4d3aba9-kube-api-access-pmjzm\") on node \"crc\" DevicePath \"\"" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.969521 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" event={"ID":"fa5259c8-c545-4b45-a9c6-e0bca4d3aba9","Type":"ContainerDied","Data":"159caac13e26af8ff7e20d53114466a89549bff12419c0bf4713ed55172141e9"} Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.969577 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="159caac13e26af8ff7e20d53114466a89549bff12419c0bf4713ed55172141e9" Dec 03 09:31:53 crc kubenswrapper[4573]: I1203 09:31:53.969683 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.270435 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c9wh4"] Dec 03 09:31:57 crc kubenswrapper[4573]: E1203 09:31:57.273132 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.273373 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 09:31:57 crc kubenswrapper[4573]: E1203 09:31:57.273538 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46e2a8e9-4614-4763-979a-daa41262f708" containerName="collect-profiles" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.273654 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="46e2a8e9-4614-4763-979a-daa41262f708" containerName="collect-profiles" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.274189 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="46e2a8e9-4614-4763-979a-daa41262f708" containerName="collect-profiles" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.274357 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5259c8-c545-4b45-a9c6-e0bca4d3aba9" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.276980 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.290989 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c9wh4"] Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.425749 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-utilities\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.426537 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-catalog-content\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.426724 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjrg4\" (UniqueName: \"kubernetes.io/projected/0320e283-e0cb-47bb-9002-0aa21119b746-kube-api-access-cjrg4\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.529115 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-catalog-content\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.529180 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjrg4\" (UniqueName: \"kubernetes.io/projected/0320e283-e0cb-47bb-9002-0aa21119b746-kube-api-access-cjrg4\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.529292 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-utilities\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.529961 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-utilities\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.529957 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-catalog-content\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.554315 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjrg4\" (UniqueName: \"kubernetes.io/projected/0320e283-e0cb-47bb-9002-0aa21119b746-kube-api-access-cjrg4\") pod \"community-operators-c9wh4\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:57 crc kubenswrapper[4573]: I1203 09:31:57.617865 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:31:58 crc kubenswrapper[4573]: I1203 09:31:58.293000 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c9wh4"] Dec 03 09:31:59 crc kubenswrapper[4573]: I1203 09:31:59.026269 4573 generic.go:334] "Generic (PLEG): container finished" podID="0320e283-e0cb-47bb-9002-0aa21119b746" containerID="d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d" exitCode=0 Dec 03 09:31:59 crc kubenswrapper[4573]: I1203 09:31:59.026341 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9wh4" event={"ID":"0320e283-e0cb-47bb-9002-0aa21119b746","Type":"ContainerDied","Data":"d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d"} Dec 03 09:31:59 crc kubenswrapper[4573]: I1203 09:31:59.026667 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9wh4" event={"ID":"0320e283-e0cb-47bb-9002-0aa21119b746","Type":"ContainerStarted","Data":"265998f251b98d73dae4da4d11ed51025d0887e91844a7993df5f6cd0890b24f"} Dec 03 09:31:59 crc kubenswrapper[4573]: I1203 09:31:59.032411 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:31:59 crc kubenswrapper[4573]: E1203 09:31:59.032665 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:32:00 crc kubenswrapper[4573]: I1203 09:32:00.073022 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9wh4" event={"ID":"0320e283-e0cb-47bb-9002-0aa21119b746","Type":"ContainerStarted","Data":"9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06"} Dec 03 09:32:01 crc kubenswrapper[4573]: I1203 09:32:01.080902 4573 generic.go:334] "Generic (PLEG): container finished" podID="0320e283-e0cb-47bb-9002-0aa21119b746" containerID="9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06" exitCode=0 Dec 03 09:32:01 crc kubenswrapper[4573]: I1203 09:32:01.080957 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9wh4" event={"ID":"0320e283-e0cb-47bb-9002-0aa21119b746","Type":"ContainerDied","Data":"9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06"} Dec 03 09:32:02 crc kubenswrapper[4573]: I1203 09:32:02.096089 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9wh4" event={"ID":"0320e283-e0cb-47bb-9002-0aa21119b746","Type":"ContainerStarted","Data":"8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8"} Dec 03 09:32:02 crc kubenswrapper[4573]: I1203 09:32:02.134555 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c9wh4" podStartSLOduration=2.614044497 podStartE2EDuration="5.134528926s" podCreationTimestamp="2025-12-03 09:31:57 +0000 UTC" firstStartedPulling="2025-12-03 09:31:59.031554005 +0000 UTC m=+3239.599933294" lastFinishedPulling="2025-12-03 09:32:01.552038424 +0000 UTC m=+3242.120417723" observedRunningTime="2025-12-03 09:32:02.122983601 +0000 UTC m=+3242.691362890" watchObservedRunningTime="2025-12-03 09:32:02.134528926 +0000 UTC m=+3242.702908195" Dec 03 09:32:07 crc kubenswrapper[4573]: I1203 09:32:07.618563 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:32:07 crc kubenswrapper[4573]: I1203 09:32:07.619319 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:32:07 crc kubenswrapper[4573]: I1203 09:32:07.689001 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:32:08 crc kubenswrapper[4573]: I1203 09:32:08.241522 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:32:08 crc kubenswrapper[4573]: I1203 09:32:08.312548 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c9wh4"] Dec 03 09:32:10 crc kubenswrapper[4573]: I1203 09:32:10.208996 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c9wh4" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" containerName="registry-server" containerID="cri-o://8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8" gracePeriod=2 Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.811633 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.826414 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjrg4\" (UniqueName: \"kubernetes.io/projected/0320e283-e0cb-47bb-9002-0aa21119b746-kube-api-access-cjrg4\") pod \"0320e283-e0cb-47bb-9002-0aa21119b746\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.826526 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-catalog-content\") pod \"0320e283-e0cb-47bb-9002-0aa21119b746\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.826592 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-utilities\") pod \"0320e283-e0cb-47bb-9002-0aa21119b746\" (UID: \"0320e283-e0cb-47bb-9002-0aa21119b746\") " Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.827799 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-utilities" (OuterVolumeSpecName: "utilities") pod "0320e283-e0cb-47bb-9002-0aa21119b746" (UID: "0320e283-e0cb-47bb-9002-0aa21119b746"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.842687 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0320e283-e0cb-47bb-9002-0aa21119b746-kube-api-access-cjrg4" (OuterVolumeSpecName: "kube-api-access-cjrg4") pod "0320e283-e0cb-47bb-9002-0aa21119b746" (UID: "0320e283-e0cb-47bb-9002-0aa21119b746"). InnerVolumeSpecName "kube-api-access-cjrg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.908610 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0320e283-e0cb-47bb-9002-0aa21119b746" (UID: "0320e283-e0cb-47bb-9002-0aa21119b746"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.928846 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.928898 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjrg4\" (UniqueName: \"kubernetes.io/projected/0320e283-e0cb-47bb-9002-0aa21119b746-kube-api-access-cjrg4\") on node \"crc\" DevicePath \"\"" Dec 03 09:32:11 crc kubenswrapper[4573]: I1203 09:32:11.928910 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0320e283-e0cb-47bb-9002-0aa21119b746-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.233135 4573 generic.go:334] "Generic (PLEG): container finished" podID="0320e283-e0cb-47bb-9002-0aa21119b746" containerID="8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8" exitCode=0 Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.233185 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9wh4" event={"ID":"0320e283-e0cb-47bb-9002-0aa21119b746","Type":"ContainerDied","Data":"8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8"} Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.233213 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c9wh4" event={"ID":"0320e283-e0cb-47bb-9002-0aa21119b746","Type":"ContainerDied","Data":"265998f251b98d73dae4da4d11ed51025d0887e91844a7993df5f6cd0890b24f"} Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.233252 4573 scope.go:117] "RemoveContainer" containerID="8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.233395 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c9wh4" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.268869 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c9wh4"] Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.275926 4573 scope.go:117] "RemoveContainer" containerID="9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.278768 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c9wh4"] Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.298986 4573 scope.go:117] "RemoveContainer" containerID="d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.366176 4573 scope.go:117] "RemoveContainer" containerID="8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8" Dec 03 09:32:12 crc kubenswrapper[4573]: E1203 09:32:12.366843 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8\": container with ID starting with 8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8 not found: ID does not exist" containerID="8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.366890 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8"} err="failed to get container status \"8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8\": rpc error: code = NotFound desc = could not find container \"8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8\": container with ID starting with 8024c546c9f815815524125c088c4f65ce8b8fd4908e412f0a6c563257b567f8 not found: ID does not exist" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.366922 4573 scope.go:117] "RemoveContainer" containerID="9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06" Dec 03 09:32:12 crc kubenswrapper[4573]: E1203 09:32:12.367578 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06\": container with ID starting with 9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06 not found: ID does not exist" containerID="9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.367619 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06"} err="failed to get container status \"9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06\": rpc error: code = NotFound desc = could not find container \"9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06\": container with ID starting with 9846e1d0c0265b985b125ffab8873dcafb17f8df61e2e5ad53809bf542646f06 not found: ID does not exist" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.367639 4573 scope.go:117] "RemoveContainer" containerID="d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d" Dec 03 09:32:12 crc kubenswrapper[4573]: E1203 09:32:12.368282 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d\": container with ID starting with d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d not found: ID does not exist" containerID="d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d" Dec 03 09:32:12 crc kubenswrapper[4573]: I1203 09:32:12.368337 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d"} err="failed to get container status \"d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d\": rpc error: code = NotFound desc = could not find container \"d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d\": container with ID starting with d05bbb77b62d37f8fcca12e3c59826f8570858f9f8c7cab813232d7781020f0d not found: ID does not exist" Dec 03 09:32:14 crc kubenswrapper[4573]: I1203 09:32:14.035607 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:32:14 crc kubenswrapper[4573]: E1203 09:32:14.037257 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:32:14 crc kubenswrapper[4573]: I1203 09:32:14.076815 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" path="/var/lib/kubelet/pods/0320e283-e0cb-47bb-9002-0aa21119b746/volumes" Dec 03 09:32:26 crc kubenswrapper[4573]: I1203 09:32:26.031247 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:32:26 crc kubenswrapper[4573]: E1203 09:32:26.032384 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:32:39 crc kubenswrapper[4573]: I1203 09:32:39.031141 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:32:39 crc kubenswrapper[4573]: E1203 09:32:39.031976 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:32:50 crc kubenswrapper[4573]: I1203 09:32:50.038370 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:32:50 crc kubenswrapper[4573]: E1203 09:32:50.041417 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.405945 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 09:32:56 crc kubenswrapper[4573]: E1203 09:32:56.408267 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" containerName="extract-content" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.408363 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" containerName="extract-content" Dec 03 09:32:56 crc kubenswrapper[4573]: E1203 09:32:56.408435 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" containerName="extract-utilities" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.408493 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" containerName="extract-utilities" Dec 03 09:32:56 crc kubenswrapper[4573]: E1203 09:32:56.408580 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" containerName="registry-server" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.408638 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" containerName="registry-server" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.408921 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="0320e283-e0cb-47bb-9002-0aa21119b746" containerName="registry-server" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.409741 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.412822 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.412857 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.413368 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.413530 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pfxr8" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.439214 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.550937 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.551094 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.551168 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87llw\" (UniqueName: \"kubernetes.io/projected/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-kube-api-access-87llw\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.551189 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.551218 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.551255 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.551273 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.551292 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.551314 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.652800 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.653187 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.653309 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87llw\" (UniqueName: \"kubernetes.io/projected/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-kube-api-access-87llw\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.653392 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.653475 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.653727 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.653806 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.653882 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.653962 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.655292 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.655691 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.656232 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.656494 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.656966 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.679918 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.682647 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.683502 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87llw\" (UniqueName: \"kubernetes.io/projected/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-kube-api-access-87llw\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.709744 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.733492 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " pod="openstack/tempest-tests-tempest" Dec 03 09:32:56 crc kubenswrapper[4573]: I1203 09:32:56.744300 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 09:32:57 crc kubenswrapper[4573]: I1203 09:32:57.297506 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 09:32:57 crc kubenswrapper[4573]: I1203 09:32:57.310963 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:32:57 crc kubenswrapper[4573]: I1203 09:32:57.758064 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b","Type":"ContainerStarted","Data":"74a544f0bb525fb50578f612a043af1f5dc0ebed8522ce58ed8adc1f5b86729f"} Dec 03 09:33:02 crc kubenswrapper[4573]: I1203 09:33:02.031126 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:33:02 crc kubenswrapper[4573]: E1203 09:33:02.031994 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:33:16 crc kubenswrapper[4573]: I1203 09:33:16.031514 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:33:16 crc kubenswrapper[4573]: E1203 09:33:16.032622 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:33:28 crc kubenswrapper[4573]: I1203 09:33:28.031386 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:33:28 crc kubenswrapper[4573]: E1203 09:33:28.032405 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:33:39 crc kubenswrapper[4573]: E1203 09:33:39.570003 4573 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 03 09:33:39 crc kubenswrapper[4573]: E1203 09:33:39.571441 4573 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-87llw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 09:33:39 crc kubenswrapper[4573]: E1203 09:33:39.572719 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" Dec 03 09:33:40 crc kubenswrapper[4573]: I1203 09:33:40.040969 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:33:40 crc kubenswrapper[4573]: E1203 09:33:40.041445 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:33:40 crc kubenswrapper[4573]: E1203 09:33:40.257985 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" Dec 03 09:33:51 crc kubenswrapper[4573]: I1203 09:33:51.030908 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:33:51 crc kubenswrapper[4573]: E1203 09:33:51.033298 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:33:54 crc kubenswrapper[4573]: I1203 09:33:54.545118 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 09:33:56 crc kubenswrapper[4573]: I1203 09:33:56.464389 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b","Type":"ContainerStarted","Data":"f0d81a02756587a09e72e1565d897fa9b12547cbeb84e52b823b02f16aee49c6"} Dec 03 09:33:56 crc kubenswrapper[4573]: I1203 09:33:56.493653 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.263314972 podStartE2EDuration="1m1.493606513s" podCreationTimestamp="2025-12-03 09:32:55 +0000 UTC" firstStartedPulling="2025-12-03 09:32:57.310591136 +0000 UTC m=+3297.878970395" lastFinishedPulling="2025-12-03 09:33:54.540882657 +0000 UTC m=+3355.109261936" observedRunningTime="2025-12-03 09:33:56.486731975 +0000 UTC m=+3357.055111234" watchObservedRunningTime="2025-12-03 09:33:56.493606513 +0000 UTC m=+3357.061985772" Dec 03 09:34:03 crc kubenswrapper[4573]: I1203 09:34:03.031436 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:34:03 crc kubenswrapper[4573]: E1203 09:34:03.032450 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:34:15 crc kubenswrapper[4573]: I1203 09:34:15.030025 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:34:15 crc kubenswrapper[4573]: E1203 09:34:15.031214 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:34:26 crc kubenswrapper[4573]: I1203 09:34:26.030870 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:34:26 crc kubenswrapper[4573]: E1203 09:34:26.031820 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:34:39 crc kubenswrapper[4573]: I1203 09:34:39.031790 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:34:39 crc kubenswrapper[4573]: E1203 09:34:39.032774 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:34:53 crc kubenswrapper[4573]: I1203 09:34:53.031625 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:34:53 crc kubenswrapper[4573]: E1203 09:34:53.032339 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:35:07 crc kubenswrapper[4573]: I1203 09:35:07.032031 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:35:07 crc kubenswrapper[4573]: I1203 09:35:07.317999 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"7a65307e8a8e0275365a4f0febbd05b2b6a7fc1f437ca3660ea4d41d4bab15c7"} Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.637767 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6z7n6"] Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.643705 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.669276 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6z7n6"] Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.705996 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nczx\" (UniqueName: \"kubernetes.io/projected/c7f3b077-a3de-4029-9c45-544774630443-kube-api-access-8nczx\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.706074 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-catalog-content\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.706098 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-utilities\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.807541 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nczx\" (UniqueName: \"kubernetes.io/projected/c7f3b077-a3de-4029-9c45-544774630443-kube-api-access-8nczx\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.807600 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-catalog-content\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.807626 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-utilities\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.808599 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-utilities\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.808889 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-catalog-content\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.837119 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nczx\" (UniqueName: \"kubernetes.io/projected/c7f3b077-a3de-4029-9c45-544774630443-kube-api-access-8nczx\") pod \"certified-operators-6z7n6\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:30 crc kubenswrapper[4573]: I1203 09:36:30.970070 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:32 crc kubenswrapper[4573]: I1203 09:36:32.241315 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6z7n6"] Dec 03 09:36:32 crc kubenswrapper[4573]: I1203 09:36:32.330227 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6z7n6" event={"ID":"c7f3b077-a3de-4029-9c45-544774630443","Type":"ContainerStarted","Data":"79ef84811fcd824a79148ceda4d8df4ffb30919dec049ed7e2bf2687f09f6165"} Dec 03 09:36:33 crc kubenswrapper[4573]: I1203 09:36:33.340406 4573 generic.go:334] "Generic (PLEG): container finished" podID="c7f3b077-a3de-4029-9c45-544774630443" containerID="01430cf98acc8d60515faadebd4f35c860a74839ad295a57b82428ffe3454941" exitCode=0 Dec 03 09:36:33 crc kubenswrapper[4573]: I1203 09:36:33.340469 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6z7n6" event={"ID":"c7f3b077-a3de-4029-9c45-544774630443","Type":"ContainerDied","Data":"01430cf98acc8d60515faadebd4f35c860a74839ad295a57b82428ffe3454941"} Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.116866 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fx4xf"] Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.122569 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.150160 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx4xf"] Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.188286 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng65q\" (UniqueName: \"kubernetes.io/projected/d5a52e2a-7e6d-464a-86dd-7141fc327a13-kube-api-access-ng65q\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.188342 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-utilities\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.188405 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-catalog-content\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.291480 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng65q\" (UniqueName: \"kubernetes.io/projected/d5a52e2a-7e6d-464a-86dd-7141fc327a13-kube-api-access-ng65q\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.291924 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-utilities\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.292259 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-catalog-content\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.293184 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-utilities\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.293219 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-catalog-content\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.316076 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng65q\" (UniqueName: \"kubernetes.io/projected/d5a52e2a-7e6d-464a-86dd-7141fc327a13-kube-api-access-ng65q\") pod \"redhat-marketplace-fx4xf\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:34 crc kubenswrapper[4573]: I1203 09:36:34.465831 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:35 crc kubenswrapper[4573]: I1203 09:36:35.179655 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx4xf"] Dec 03 09:36:35 crc kubenswrapper[4573]: I1203 09:36:35.367925 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx4xf" event={"ID":"d5a52e2a-7e6d-464a-86dd-7141fc327a13","Type":"ContainerStarted","Data":"51da809ee9a2d13b74a9225b0559e771e0c59f369e95a900b180009d40c5f99c"} Dec 03 09:36:35 crc kubenswrapper[4573]: I1203 09:36:35.371690 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6z7n6" event={"ID":"c7f3b077-a3de-4029-9c45-544774630443","Type":"ContainerStarted","Data":"5af41fc9bdc24034c8e425dc855fdb8a07909bc68de3c1b293232b53c19ad891"} Dec 03 09:36:36 crc kubenswrapper[4573]: I1203 09:36:36.382441 4573 generic.go:334] "Generic (PLEG): container finished" podID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerID="bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc" exitCode=0 Dec 03 09:36:36 crc kubenswrapper[4573]: I1203 09:36:36.382548 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx4xf" event={"ID":"d5a52e2a-7e6d-464a-86dd-7141fc327a13","Type":"ContainerDied","Data":"bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc"} Dec 03 09:36:36 crc kubenswrapper[4573]: I1203 09:36:36.386210 4573 generic.go:334] "Generic (PLEG): container finished" podID="c7f3b077-a3de-4029-9c45-544774630443" containerID="5af41fc9bdc24034c8e425dc855fdb8a07909bc68de3c1b293232b53c19ad891" exitCode=0 Dec 03 09:36:36 crc kubenswrapper[4573]: I1203 09:36:36.386276 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6z7n6" event={"ID":"c7f3b077-a3de-4029-9c45-544774630443","Type":"ContainerDied","Data":"5af41fc9bdc24034c8e425dc855fdb8a07909bc68de3c1b293232b53c19ad891"} Dec 03 09:36:37 crc kubenswrapper[4573]: I1203 09:36:37.409217 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6z7n6" event={"ID":"c7f3b077-a3de-4029-9c45-544774630443","Type":"ContainerStarted","Data":"bf8cab3a96df7f69d401eb8fbc07581cb8fb5cff322d57b2d6205a4a415c4106"} Dec 03 09:36:37 crc kubenswrapper[4573]: I1203 09:36:37.441202 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6z7n6" podStartSLOduration=3.976590164 podStartE2EDuration="7.441176259s" podCreationTimestamp="2025-12-03 09:36:30 +0000 UTC" firstStartedPulling="2025-12-03 09:36:33.342259073 +0000 UTC m=+3513.910638322" lastFinishedPulling="2025-12-03 09:36:36.806845158 +0000 UTC m=+3517.375224417" observedRunningTime="2025-12-03 09:36:37.435579296 +0000 UTC m=+3518.003958555" watchObservedRunningTime="2025-12-03 09:36:37.441176259 +0000 UTC m=+3518.009555518" Dec 03 09:36:38 crc kubenswrapper[4573]: I1203 09:36:38.423823 4573 generic.go:334] "Generic (PLEG): container finished" podID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerID="85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99" exitCode=0 Dec 03 09:36:38 crc kubenswrapper[4573]: I1203 09:36:38.425617 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx4xf" event={"ID":"d5a52e2a-7e6d-464a-86dd-7141fc327a13","Type":"ContainerDied","Data":"85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99"} Dec 03 09:36:39 crc kubenswrapper[4573]: I1203 09:36:39.438725 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx4xf" event={"ID":"d5a52e2a-7e6d-464a-86dd-7141fc327a13","Type":"ContainerStarted","Data":"87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747"} Dec 03 09:36:39 crc kubenswrapper[4573]: I1203 09:36:39.462314 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fx4xf" podStartSLOduration=3.992512859 podStartE2EDuration="6.462286678s" podCreationTimestamp="2025-12-03 09:36:33 +0000 UTC" firstStartedPulling="2025-12-03 09:36:36.385168914 +0000 UTC m=+3516.953548173" lastFinishedPulling="2025-12-03 09:36:38.854942723 +0000 UTC m=+3519.423321992" observedRunningTime="2025-12-03 09:36:39.460021556 +0000 UTC m=+3520.028400815" watchObservedRunningTime="2025-12-03 09:36:39.462286678 +0000 UTC m=+3520.030665937" Dec 03 09:36:40 crc kubenswrapper[4573]: I1203 09:36:40.970471 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:40 crc kubenswrapper[4573]: I1203 09:36:40.971934 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:42 crc kubenswrapper[4573]: I1203 09:36:42.069013 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-6z7n6" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="registry-server" probeResult="failure" output=< Dec 03 09:36:42 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:36:42 crc kubenswrapper[4573]: > Dec 03 09:36:44 crc kubenswrapper[4573]: I1203 09:36:44.466520 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:44 crc kubenswrapper[4573]: I1203 09:36:44.467923 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:44 crc kubenswrapper[4573]: I1203 09:36:44.522519 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:45 crc kubenswrapper[4573]: I1203 09:36:45.547030 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:45 crc kubenswrapper[4573]: I1203 09:36:45.621562 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx4xf"] Dec 03 09:36:47 crc kubenswrapper[4573]: I1203 09:36:47.583590 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fx4xf" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerName="registry-server" containerID="cri-o://87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747" gracePeriod=2 Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.274183 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.387955 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng65q\" (UniqueName: \"kubernetes.io/projected/d5a52e2a-7e6d-464a-86dd-7141fc327a13-kube-api-access-ng65q\") pod \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.388034 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-utilities\") pod \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.388107 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-catalog-content\") pod \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\" (UID: \"d5a52e2a-7e6d-464a-86dd-7141fc327a13\") " Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.389901 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-utilities" (OuterVolumeSpecName: "utilities") pod "d5a52e2a-7e6d-464a-86dd-7141fc327a13" (UID: "d5a52e2a-7e6d-464a-86dd-7141fc327a13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.411128 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5a52e2a-7e6d-464a-86dd-7141fc327a13-kube-api-access-ng65q" (OuterVolumeSpecName: "kube-api-access-ng65q") pod "d5a52e2a-7e6d-464a-86dd-7141fc327a13" (UID: "d5a52e2a-7e6d-464a-86dd-7141fc327a13"). InnerVolumeSpecName "kube-api-access-ng65q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.417890 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5a52e2a-7e6d-464a-86dd-7141fc327a13" (UID: "d5a52e2a-7e6d-464a-86dd-7141fc327a13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.491290 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng65q\" (UniqueName: \"kubernetes.io/projected/d5a52e2a-7e6d-464a-86dd-7141fc327a13-kube-api-access-ng65q\") on node \"crc\" DevicePath \"\"" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.491342 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.491356 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5a52e2a-7e6d-464a-86dd-7141fc327a13-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.596436 4573 generic.go:334] "Generic (PLEG): container finished" podID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerID="87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747" exitCode=0 Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.596485 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx4xf" event={"ID":"d5a52e2a-7e6d-464a-86dd-7141fc327a13","Type":"ContainerDied","Data":"87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747"} Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.596517 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fx4xf" event={"ID":"d5a52e2a-7e6d-464a-86dd-7141fc327a13","Type":"ContainerDied","Data":"51da809ee9a2d13b74a9225b0559e771e0c59f369e95a900b180009d40c5f99c"} Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.596541 4573 scope.go:117] "RemoveContainer" containerID="87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.596704 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fx4xf" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.633337 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx4xf"] Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.639511 4573 scope.go:117] "RemoveContainer" containerID="85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.644287 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fx4xf"] Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.665191 4573 scope.go:117] "RemoveContainer" containerID="bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.719652 4573 scope.go:117] "RemoveContainer" containerID="87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747" Dec 03 09:36:48 crc kubenswrapper[4573]: E1203 09:36:48.720419 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747\": container with ID starting with 87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747 not found: ID does not exist" containerID="87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.720539 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747"} err="failed to get container status \"87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747\": rpc error: code = NotFound desc = could not find container \"87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747\": container with ID starting with 87390c0537b4a78d892bc74bb2e431c80c218722f458cda53c386cb4c180c747 not found: ID does not exist" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.720583 4573 scope.go:117] "RemoveContainer" containerID="85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99" Dec 03 09:36:48 crc kubenswrapper[4573]: E1203 09:36:48.721104 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99\": container with ID starting with 85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99 not found: ID does not exist" containerID="85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.721152 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99"} err="failed to get container status \"85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99\": rpc error: code = NotFound desc = could not find container \"85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99\": container with ID starting with 85be8a148fb69aad926cd5ed45538353a28ee0142663456d513e019316467f99 not found: ID does not exist" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.721171 4573 scope.go:117] "RemoveContainer" containerID="bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc" Dec 03 09:36:48 crc kubenswrapper[4573]: E1203 09:36:48.721715 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc\": container with ID starting with bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc not found: ID does not exist" containerID="bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc" Dec 03 09:36:48 crc kubenswrapper[4573]: I1203 09:36:48.721769 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc"} err="failed to get container status \"bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc\": rpc error: code = NotFound desc = could not find container \"bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc\": container with ID starting with bf3997a5bbfbc2d7900947371cbbfecdc3147445dcbe02217a10a440d4b18bfc not found: ID does not exist" Dec 03 09:36:50 crc kubenswrapper[4573]: I1203 09:36:50.056571 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" path="/var/lib/kubelet/pods/d5a52e2a-7e6d-464a-86dd-7141fc327a13/volumes" Dec 03 09:36:51 crc kubenswrapper[4573]: I1203 09:36:51.047125 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:51 crc kubenswrapper[4573]: I1203 09:36:51.110170 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:51 crc kubenswrapper[4573]: I1203 09:36:51.514707 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6z7n6"] Dec 03 09:36:52 crc kubenswrapper[4573]: I1203 09:36:52.639504 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6z7n6" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="registry-server" containerID="cri-o://bf8cab3a96df7f69d401eb8fbc07581cb8fb5cff322d57b2d6205a4a415c4106" gracePeriod=2 Dec 03 09:36:53 crc kubenswrapper[4573]: I1203 09:36:53.695193 4573 generic.go:334] "Generic (PLEG): container finished" podID="c7f3b077-a3de-4029-9c45-544774630443" containerID="bf8cab3a96df7f69d401eb8fbc07581cb8fb5cff322d57b2d6205a4a415c4106" exitCode=0 Dec 03 09:36:53 crc kubenswrapper[4573]: I1203 09:36:53.695550 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6z7n6" event={"ID":"c7f3b077-a3de-4029-9c45-544774630443","Type":"ContainerDied","Data":"bf8cab3a96df7f69d401eb8fbc07581cb8fb5cff322d57b2d6205a4a415c4106"} Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.263456 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.430169 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-catalog-content\") pod \"c7f3b077-a3de-4029-9c45-544774630443\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.433283 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nczx\" (UniqueName: \"kubernetes.io/projected/c7f3b077-a3de-4029-9c45-544774630443-kube-api-access-8nczx\") pod \"c7f3b077-a3de-4029-9c45-544774630443\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.437953 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-utilities\") pod \"c7f3b077-a3de-4029-9c45-544774630443\" (UID: \"c7f3b077-a3de-4029-9c45-544774630443\") " Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.438958 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-utilities" (OuterVolumeSpecName: "utilities") pod "c7f3b077-a3de-4029-9c45-544774630443" (UID: "c7f3b077-a3de-4029-9c45-544774630443"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.442690 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7f3b077-a3de-4029-9c45-544774630443-kube-api-access-8nczx" (OuterVolumeSpecName: "kube-api-access-8nczx") pod "c7f3b077-a3de-4029-9c45-544774630443" (UID: "c7f3b077-a3de-4029-9c45-544774630443"). InnerVolumeSpecName "kube-api-access-8nczx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.453690 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.503580 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7f3b077-a3de-4029-9c45-544774630443" (UID: "c7f3b077-a3de-4029-9c45-544774630443"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.555270 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nczx\" (UniqueName: \"kubernetes.io/projected/c7f3b077-a3de-4029-9c45-544774630443-kube-api-access-8nczx\") on node \"crc\" DevicePath \"\"" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.555307 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7f3b077-a3de-4029-9c45-544774630443-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.707194 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6z7n6" event={"ID":"c7f3b077-a3de-4029-9c45-544774630443","Type":"ContainerDied","Data":"79ef84811fcd824a79148ceda4d8df4ffb30919dec049ed7e2bf2687f09f6165"} Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.707256 4573 scope.go:117] "RemoveContainer" containerID="bf8cab3a96df7f69d401eb8fbc07581cb8fb5cff322d57b2d6205a4a415c4106" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.709198 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6z7n6" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.755127 4573 scope.go:117] "RemoveContainer" containerID="5af41fc9bdc24034c8e425dc855fdb8a07909bc68de3c1b293232b53c19ad891" Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.765465 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6z7n6"] Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.782757 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6z7n6"] Dec 03 09:36:54 crc kubenswrapper[4573]: I1203 09:36:54.794956 4573 scope.go:117] "RemoveContainer" containerID="01430cf98acc8d60515faadebd4f35c860a74839ad295a57b82428ffe3454941" Dec 03 09:36:56 crc kubenswrapper[4573]: I1203 09:36:56.050757 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7f3b077-a3de-4029-9c45-544774630443" path="/var/lib/kubelet/pods/c7f3b077-a3de-4029-9c45-544774630443/volumes" Dec 03 09:37:26 crc kubenswrapper[4573]: I1203 09:37:26.943070 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:37:26 crc kubenswrapper[4573]: I1203 09:37:26.943851 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.157125 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f9gmr"] Dec 03 09:37:44 crc kubenswrapper[4573]: E1203 09:37:44.158673 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="extract-content" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.158705 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="extract-content" Dec 03 09:37:44 crc kubenswrapper[4573]: E1203 09:37:44.158729 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerName="extract-content" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.158737 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerName="extract-content" Dec 03 09:37:44 crc kubenswrapper[4573]: E1203 09:37:44.158754 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="registry-server" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.158760 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="registry-server" Dec 03 09:37:44 crc kubenswrapper[4573]: E1203 09:37:44.158791 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerName="extract-utilities" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.158800 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerName="extract-utilities" Dec 03 09:37:44 crc kubenswrapper[4573]: E1203 09:37:44.158817 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerName="registry-server" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.158823 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerName="registry-server" Dec 03 09:37:44 crc kubenswrapper[4573]: E1203 09:37:44.158843 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="extract-utilities" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.158850 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="extract-utilities" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.159118 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7f3b077-a3de-4029-9c45-544774630443" containerName="registry-server" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.159144 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5a52e2a-7e6d-464a-86dd-7141fc327a13" containerName="registry-server" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.161205 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.168163 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9gmr"] Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.199186 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzfts\" (UniqueName: \"kubernetes.io/projected/be95fc26-9875-4a19-acbd-4be7fc4d3224-kube-api-access-dzfts\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.199534 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-utilities\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.199741 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-catalog-content\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.302910 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzfts\" (UniqueName: \"kubernetes.io/projected/be95fc26-9875-4a19-acbd-4be7fc4d3224-kube-api-access-dzfts\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.303027 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-utilities\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.303276 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-catalog-content\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.304089 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-catalog-content\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.304150 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-utilities\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.326685 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzfts\" (UniqueName: \"kubernetes.io/projected/be95fc26-9875-4a19-acbd-4be7fc4d3224-kube-api-access-dzfts\") pod \"redhat-operators-f9gmr\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:44 crc kubenswrapper[4573]: I1203 09:37:44.494357 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:37:45 crc kubenswrapper[4573]: I1203 09:37:45.120514 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9gmr"] Dec 03 09:37:45 crc kubenswrapper[4573]: I1203 09:37:45.234477 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9gmr" event={"ID":"be95fc26-9875-4a19-acbd-4be7fc4d3224","Type":"ContainerStarted","Data":"fdc16a080bd25dcbf3b3068e559802131b917ed6b7839699862bce4ab2d5fc21"} Dec 03 09:37:46 crc kubenswrapper[4573]: I1203 09:37:46.246578 4573 generic.go:334] "Generic (PLEG): container finished" podID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerID="a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6" exitCode=0 Dec 03 09:37:46 crc kubenswrapper[4573]: I1203 09:37:46.246685 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9gmr" event={"ID":"be95fc26-9875-4a19-acbd-4be7fc4d3224","Type":"ContainerDied","Data":"a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6"} Dec 03 09:37:50 crc kubenswrapper[4573]: I1203 09:37:50.340810 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9gmr" event={"ID":"be95fc26-9875-4a19-acbd-4be7fc4d3224","Type":"ContainerStarted","Data":"fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c"} Dec 03 09:37:56 crc kubenswrapper[4573]: I1203 09:37:56.722609 4573 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-l7qrz container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 09:37:56 crc kubenswrapper[4573]: I1203 09:37:56.725316 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l7qrz" podUID="8b5ee846-009f-4feb-a61b-a413aee18c4d" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:37:56 crc kubenswrapper[4573]: I1203 09:37:56.943196 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:37:56 crc kubenswrapper[4573]: I1203 09:37:56.943302 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:37:58 crc kubenswrapper[4573]: I1203 09:37:58.733423 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovsdbserver-nb-0" podUID="1604082b-b77f-4df2-bea3-ab7803b3e5ec" containerName="ovsdbserver-nb" probeResult="failure" output="command timed out" Dec 03 09:37:58 crc kubenswrapper[4573]: I1203 09:37:58.733462 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="56f569d8-50e2-4177-bec8-8225de477357" containerName="nova-scheduler-scheduler" probeResult="failure" output="command timed out" Dec 03 09:37:58 crc kubenswrapper[4573]: I1203 09:37:58.806341 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="3f75b291-5f0e-422f-a640-a3a231c400ee" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:37:58 crc kubenswrapper[4573]: I1203 09:37:58.806359 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="3f75b291-5f0e-422f-a640-a3a231c400ee" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:37:59 crc kubenswrapper[4573]: I1203 09:37:59.732318 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="3601284e-acb1-4e60-a4c1-18c404dc4072" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 03 09:38:01 crc kubenswrapper[4573]: I1203 09:38:01.712198 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-6fjjm" podUID="2d98f6ef-9108-408a-8ae6-b1116c434d68" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:38:01 crc kubenswrapper[4573]: I1203 09:38:01.743778 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="1604082b-b77f-4df2-bea3-ab7803b3e5ec" containerName="ovsdbserver-nb" probeResult="failure" output="command timed out" Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.473812 4573 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 5.740778847s: [/var/lib/containers/storage/overlay/343c9cc65882e4828cc9f1df7f4602834678d1dc623f07d3517b6e104e833bb8/diff /var/log/pods/openstack_ovsdbserver-sb-0_f5af474c-26c9-4877-989a-911cd77862ab/openstack-network-exporter/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.475518 4573 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.734612311s: [/var/lib/containers/storage/overlay/b9b386280ce57621b2cadd5c06ebcbe5cad2a8d313880b56c9a0c6534c96ce22/diff /var/log/pods/openstack_neutron-6cfccd4555-2shp6_4434273d-ce2e-4ecd-820d-7f2ecf35bca7/neutron-httpd/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.482183 4573 trace.go:236] Trace[1543821516]: "Calculate volume metrics of scripts for pod openstack/ovn-controller-ovs-jbx79" (03-Dec-2025 09:37:54.771) (total time: 8708ms): Dec 03 09:38:03 crc kubenswrapper[4573]: Trace[1543821516]: [8.708082139s] [8.708082139s] END Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.485979 4573 trace.go:236] Trace[79466051]: "Calculate volume metrics of test-operator-ephemeral-temporary for pod openstack/tempest-tests-tempest" (03-Dec-2025 09:37:59.008) (total time: 4477ms): Dec 03 09:38:03 crc kubenswrapper[4573]: Trace[79466051]: [4.477363961s] [4.477363961s] END Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.491649 4573 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 3.594687291s: [/var/lib/containers/storage/overlay/4ff46ebd3ce717dcab0255002a867d4ef6dc1b287ba3a8afa4dbe6ce89cc5b22/diff /var/log/pods/openshift-authentication_oauth-openshift-7b4b9565b9-sjwxg_7da099a6-df67-490b-b5ab-2cb6cc44e438/oauth-openshift/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.491750 4573 trace.go:236] Trace[191104797]: "Calculate volume metrics of httpd-run for pod openstack/glance-default-external-api-0" (03-Dec-2025 09:37:59.241) (total time: 4249ms): Dec 03 09:38:03 crc kubenswrapper[4573]: Trace[191104797]: [4.249763017s] [4.249763017s] END Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.493217 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96" containerName="galera" probeResult="failure" output="command timed out" Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.493739 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96" containerName="galera" probeResult="failure" output="command timed out" Dec 03 09:38:03 crc kubenswrapper[4573]: E1203 09:38:03.497949 4573 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="9.468s" Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.512616 4573 trace.go:236] Trace[899678737]: "Calculate volume metrics of scripts for pod openstack/ovn-controller-bj74b" (03-Dec-2025 09:37:53.069) (total time: 10442ms): Dec 03 09:38:03 crc kubenswrapper[4573]: Trace[899678737]: [10.442858723s] [10.442858723s] END Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.513666 4573 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 9.186275473s: [/var/lib/containers/storage/overlay/3e39090685130b30d05b0f1b964d161da0d7509fd244e83f6249c3ad40514832/diff /var/log/pods/openstack_nova-scheduler-0_56f569d8-50e2-4177-bec8-8225de477357/nova-scheduler-scheduler/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.523853 4573 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 10.158758948s: [/var/lib/containers/storage/overlay/e2ce75263d6c2b691b89d64f112ad66279319d5f0da30fc432dda7c230a554ce/diff /var/log/pods/openstack_openstack-galera-0_6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96/galera/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.545134 4573 patch_prober.go:28] interesting pod/controller-manager-99456b778-8tw6l container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.545198 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-99456b778-8tw6l" podUID="84a8db50-4c2a-40eb-ad99-1e19c9cc4455" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.549907 4573 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-z2tsl container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.586288 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-z2tsl" podUID="f6e10744-6f93-416a-a6aa-4a7a07c76eff" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.564651 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="20097614-b580-49bf-a8c7-04f371825b9a" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.167:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.703195 4573 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 10.668829601s: [/var/lib/containers/storage/overlay/7f358748dfd251133925f543ceda11d7040e59cae44f74ca6e52ac24d30e029d/diff /var/log/pods/openstack_openstack-cell1-galera-0_9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d/galera/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.747767 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovsdbserver-nb-0" podUID="1604082b-b77f-4df2-bea3-ab7803b3e5ec" containerName="ovsdbserver-nb" probeResult="failure" output="command timed out" Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.809774 4573 trace.go:236] Trace[379762337]: "iptables ChainExists" (03-Dec-2025 09:38:01.672) (total time: 2136ms): Dec 03 09:38:03 crc kubenswrapper[4573]: Trace[379762337]: [2.136862406s] [2.136862406s] END Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.809842 4573 trace.go:236] Trace[901126830]: "iptables ChainExists" (03-Dec-2025 09:38:01.704) (total time: 2105ms): Dec 03 09:38:03 crc kubenswrapper[4573]: Trace[901126830]: [2.105457458s] [2.105457458s] END Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.817972 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="3f75b291-5f0e-422f-a640-a3a231c400ee" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 09:38:03 crc kubenswrapper[4573]: I1203 09:38:03.818326 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="3f75b291-5f0e-422f-a640-a3a231c400ee" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 09:38:04 crc kubenswrapper[4573]: I1203 09:38:04.003237 4573 trace.go:236] Trace[1367473004]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/redhat-operators-f9gmr" (03-Dec-2025 09:38:01.689) (total time: 2313ms): Dec 03 09:38:04 crc kubenswrapper[4573]: Trace[1367473004]: [2.313697286s] [2.313697286s] END Dec 03 09:38:04 crc kubenswrapper[4573]: I1203 09:38:04.019572 4573 trace.go:236] Trace[368588175]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/certified-operators-6zq2f" (03-Dec-2025 09:37:52.988) (total time: 11031ms): Dec 03 09:38:04 crc kubenswrapper[4573]: Trace[368588175]: [11.031387314s] [11.031387314s] END Dec 03 09:38:04 crc kubenswrapper[4573]: I1203 09:38:04.080806 4573 trace.go:236] Trace[646325234]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/redhat-operators-tmzcx" (03-Dec-2025 09:37:54.583) (total time: 9497ms): Dec 03 09:38:04 crc kubenswrapper[4573]: Trace[646325234]: [9.497441503s] [9.497441503s] END Dec 03 09:38:04 crc kubenswrapper[4573]: I1203 09:38:04.747137 4573 generic.go:334] "Generic (PLEG): container finished" podID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerID="fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c" exitCode=0 Dec 03 09:38:04 crc kubenswrapper[4573]: I1203 09:38:04.747201 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9gmr" event={"ID":"be95fc26-9875-4a19-acbd-4be7fc4d3224","Type":"ContainerDied","Data":"fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c"} Dec 03 09:38:04 crc kubenswrapper[4573]: I1203 09:38:04.757314 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:38:05 crc kubenswrapper[4573]: I1203 09:38:05.758814 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9gmr" event={"ID":"be95fc26-9875-4a19-acbd-4be7fc4d3224","Type":"ContainerStarted","Data":"97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683"} Dec 03 09:38:05 crc kubenswrapper[4573]: I1203 09:38:05.784152 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f9gmr" podStartSLOduration=2.731016475 podStartE2EDuration="21.784130028s" podCreationTimestamp="2025-12-03 09:37:44 +0000 UTC" firstStartedPulling="2025-12-03 09:37:46.24914739 +0000 UTC m=+3586.817526649" lastFinishedPulling="2025-12-03 09:38:05.302260943 +0000 UTC m=+3605.870640202" observedRunningTime="2025-12-03 09:38:05.779130661 +0000 UTC m=+3606.347509930" watchObservedRunningTime="2025-12-03 09:38:05.784130028 +0000 UTC m=+3606.352509287" Dec 03 09:38:14 crc kubenswrapper[4573]: I1203 09:38:14.494973 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:38:14 crc kubenswrapper[4573]: I1203 09:38:14.497549 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:38:15 crc kubenswrapper[4573]: I1203 09:38:15.908152 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f9gmr" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="registry-server" probeResult="failure" output=< Dec 03 09:38:15 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:38:15 crc kubenswrapper[4573]: > Dec 03 09:38:24 crc kubenswrapper[4573]: I1203 09:38:24.556123 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:38:24 crc kubenswrapper[4573]: I1203 09:38:24.626443 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:38:24 crc kubenswrapper[4573]: I1203 09:38:24.813808 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9gmr"] Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.044969 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f9gmr" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="registry-server" containerID="cri-o://97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683" gracePeriod=2 Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.622493 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.743808 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-catalog-content\") pod \"be95fc26-9875-4a19-acbd-4be7fc4d3224\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.743931 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-utilities\") pod \"be95fc26-9875-4a19-acbd-4be7fc4d3224\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.744112 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzfts\" (UniqueName: \"kubernetes.io/projected/be95fc26-9875-4a19-acbd-4be7fc4d3224-kube-api-access-dzfts\") pod \"be95fc26-9875-4a19-acbd-4be7fc4d3224\" (UID: \"be95fc26-9875-4a19-acbd-4be7fc4d3224\") " Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.745659 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-utilities" (OuterVolumeSpecName: "utilities") pod "be95fc26-9875-4a19-acbd-4be7fc4d3224" (UID: "be95fc26-9875-4a19-acbd-4be7fc4d3224"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.751946 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be95fc26-9875-4a19-acbd-4be7fc4d3224-kube-api-access-dzfts" (OuterVolumeSpecName: "kube-api-access-dzfts") pod "be95fc26-9875-4a19-acbd-4be7fc4d3224" (UID: "be95fc26-9875-4a19-acbd-4be7fc4d3224"). InnerVolumeSpecName "kube-api-access-dzfts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.846635 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzfts\" (UniqueName: \"kubernetes.io/projected/be95fc26-9875-4a19-acbd-4be7fc4d3224-kube-api-access-dzfts\") on node \"crc\" DevicePath \"\"" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.847038 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.876980 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be95fc26-9875-4a19-acbd-4be7fc4d3224" (UID: "be95fc26-9875-4a19-acbd-4be7fc4d3224"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.943558 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.943672 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.943746 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.945059 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a65307e8a8e0275365a4f0febbd05b2b6a7fc1f437ca3660ea4d41d4bab15c7"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.945136 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://7a65307e8a8e0275365a4f0febbd05b2b6a7fc1f437ca3660ea4d41d4bab15c7" gracePeriod=600 Dec 03 09:38:26 crc kubenswrapper[4573]: I1203 09:38:26.949028 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be95fc26-9875-4a19-acbd-4be7fc4d3224-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.059797 4573 generic.go:334] "Generic (PLEG): container finished" podID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerID="97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683" exitCode=0 Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.059856 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9gmr" event={"ID":"be95fc26-9875-4a19-acbd-4be7fc4d3224","Type":"ContainerDied","Data":"97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683"} Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.059898 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9gmr" event={"ID":"be95fc26-9875-4a19-acbd-4be7fc4d3224","Type":"ContainerDied","Data":"fdc16a080bd25dcbf3b3068e559802131b917ed6b7839699862bce4ab2d5fc21"} Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.059951 4573 scope.go:117] "RemoveContainer" containerID="97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.060123 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9gmr" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.129109 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9gmr"] Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.138825 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f9gmr"] Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.142510 4573 scope.go:117] "RemoveContainer" containerID="fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.186232 4573 scope.go:117] "RemoveContainer" containerID="a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.274685 4573 scope.go:117] "RemoveContainer" containerID="97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683" Dec 03 09:38:27 crc kubenswrapper[4573]: E1203 09:38:27.275318 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683\": container with ID starting with 97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683 not found: ID does not exist" containerID="97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.275397 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683"} err="failed to get container status \"97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683\": rpc error: code = NotFound desc = could not find container \"97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683\": container with ID starting with 97032bfdc14a8eff27f844d133269134e7df6e8db4681fbcf09bf35caf113683 not found: ID does not exist" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.275452 4573 scope.go:117] "RemoveContainer" containerID="fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c" Dec 03 09:38:27 crc kubenswrapper[4573]: E1203 09:38:27.275802 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c\": container with ID starting with fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c not found: ID does not exist" containerID="fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.275861 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c"} err="failed to get container status \"fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c\": rpc error: code = NotFound desc = could not find container \"fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c\": container with ID starting with fc7ab456a24791fa96083daf1e6cf4db323bd99ec24d9204374043864e23635c not found: ID does not exist" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.275905 4573 scope.go:117] "RemoveContainer" containerID="a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6" Dec 03 09:38:27 crc kubenswrapper[4573]: E1203 09:38:27.276223 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6\": container with ID starting with a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6 not found: ID does not exist" containerID="a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6" Dec 03 09:38:27 crc kubenswrapper[4573]: I1203 09:38:27.276260 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6"} err="failed to get container status \"a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6\": rpc error: code = NotFound desc = could not find container \"a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6\": container with ID starting with a7178f9192aee1698ef57603c947e6ebed614c618a95fbb90cb90aa831775de6 not found: ID does not exist" Dec 03 09:38:28 crc kubenswrapper[4573]: I1203 09:38:28.042250 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" path="/var/lib/kubelet/pods/be95fc26-9875-4a19-acbd-4be7fc4d3224/volumes" Dec 03 09:38:28 crc kubenswrapper[4573]: I1203 09:38:28.073364 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="7a65307e8a8e0275365a4f0febbd05b2b6a7fc1f437ca3660ea4d41d4bab15c7" exitCode=0 Dec 03 09:38:28 crc kubenswrapper[4573]: I1203 09:38:28.073425 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"7a65307e8a8e0275365a4f0febbd05b2b6a7fc1f437ca3660ea4d41d4bab15c7"} Dec 03 09:38:28 crc kubenswrapper[4573]: I1203 09:38:28.073481 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404"} Dec 03 09:38:28 crc kubenswrapper[4573]: I1203 09:38:28.073549 4573 scope.go:117] "RemoveContainer" containerID="b30118861f53778a1a1d74178011be89ff40beb78c01e56df92fc5bd007aeeb8" Dec 03 09:40:56 crc kubenswrapper[4573]: I1203 09:40:56.943580 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:40:56 crc kubenswrapper[4573]: I1203 09:40:56.944283 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:41:26 crc kubenswrapper[4573]: I1203 09:41:26.943838 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:41:26 crc kubenswrapper[4573]: I1203 09:41:26.944518 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:41:56 crc kubenswrapper[4573]: I1203 09:41:56.943577 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:41:56 crc kubenswrapper[4573]: I1203 09:41:56.944114 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:41:56 crc kubenswrapper[4573]: I1203 09:41:56.944168 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:41:56 crc kubenswrapper[4573]: I1203 09:41:56.944860 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:41:56 crc kubenswrapper[4573]: I1203 09:41:56.944949 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" gracePeriod=600 Dec 03 09:41:57 crc kubenswrapper[4573]: E1203 09:41:57.094322 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:41:57 crc kubenswrapper[4573]: I1203 09:41:57.647280 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" exitCode=0 Dec 03 09:41:57 crc kubenswrapper[4573]: I1203 09:41:57.647702 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404"} Dec 03 09:41:57 crc kubenswrapper[4573]: I1203 09:41:57.647761 4573 scope.go:117] "RemoveContainer" containerID="7a65307e8a8e0275365a4f0febbd05b2b6a7fc1f437ca3660ea4d41d4bab15c7" Dec 03 09:41:57 crc kubenswrapper[4573]: I1203 09:41:57.648548 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:41:57 crc kubenswrapper[4573]: E1203 09:41:57.648851 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:42:09 crc kubenswrapper[4573]: I1203 09:42:09.042945 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:42:09 crc kubenswrapper[4573]: E1203 09:42:09.044683 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.193250 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6fg8l"] Dec 03 09:42:12 crc kubenswrapper[4573]: E1203 09:42:12.194599 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="extract-content" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.194615 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="extract-content" Dec 03 09:42:12 crc kubenswrapper[4573]: E1203 09:42:12.194633 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="extract-utilities" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.194641 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="extract-utilities" Dec 03 09:42:12 crc kubenswrapper[4573]: E1203 09:42:12.194682 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="registry-server" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.194688 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="registry-server" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.194902 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="be95fc26-9875-4a19-acbd-4be7fc4d3224" containerName="registry-server" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.196306 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.224837 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6fg8l"] Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.272464 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-utilities\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.272651 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g76qn\" (UniqueName: \"kubernetes.io/projected/b529dd20-dbe0-4643-a289-8a3c9f882431-kube-api-access-g76qn\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.272701 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-catalog-content\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.375780 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-utilities\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.376021 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g76qn\" (UniqueName: \"kubernetes.io/projected/b529dd20-dbe0-4643-a289-8a3c9f882431-kube-api-access-g76qn\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.376249 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-catalog-content\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.377032 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-catalog-content\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.377495 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-utilities\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.407424 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g76qn\" (UniqueName: \"kubernetes.io/projected/b529dd20-dbe0-4643-a289-8a3c9f882431-kube-api-access-g76qn\") pod \"community-operators-6fg8l\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:12 crc kubenswrapper[4573]: I1203 09:42:12.535463 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:13 crc kubenswrapper[4573]: I1203 09:42:13.218977 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6fg8l"] Dec 03 09:42:13 crc kubenswrapper[4573]: I1203 09:42:13.881569 4573 generic.go:334] "Generic (PLEG): container finished" podID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerID="fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f" exitCode=0 Dec 03 09:42:13 crc kubenswrapper[4573]: I1203 09:42:13.881943 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fg8l" event={"ID":"b529dd20-dbe0-4643-a289-8a3c9f882431","Type":"ContainerDied","Data":"fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f"} Dec 03 09:42:13 crc kubenswrapper[4573]: I1203 09:42:13.882464 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fg8l" event={"ID":"b529dd20-dbe0-4643-a289-8a3c9f882431","Type":"ContainerStarted","Data":"c413768d0967208019897a51f7aec031563057daef2af7be5be1d158632dd6b8"} Dec 03 09:42:15 crc kubenswrapper[4573]: I1203 09:42:15.906157 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fg8l" event={"ID":"b529dd20-dbe0-4643-a289-8a3c9f882431","Type":"ContainerStarted","Data":"5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1"} Dec 03 09:42:16 crc kubenswrapper[4573]: I1203 09:42:16.927912 4573 generic.go:334] "Generic (PLEG): container finished" podID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerID="5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1" exitCode=0 Dec 03 09:42:16 crc kubenswrapper[4573]: I1203 09:42:16.927957 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fg8l" event={"ID":"b529dd20-dbe0-4643-a289-8a3c9f882431","Type":"ContainerDied","Data":"5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1"} Dec 03 09:42:17 crc kubenswrapper[4573]: I1203 09:42:17.939520 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fg8l" event={"ID":"b529dd20-dbe0-4643-a289-8a3c9f882431","Type":"ContainerStarted","Data":"b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b"} Dec 03 09:42:17 crc kubenswrapper[4573]: I1203 09:42:17.967825 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6fg8l" podStartSLOduration=2.2895458 podStartE2EDuration="5.967753055s" podCreationTimestamp="2025-12-03 09:42:12 +0000 UTC" firstStartedPulling="2025-12-03 09:42:13.883605997 +0000 UTC m=+3854.451985276" lastFinishedPulling="2025-12-03 09:42:17.561813272 +0000 UTC m=+3858.130192531" observedRunningTime="2025-12-03 09:42:17.967433026 +0000 UTC m=+3858.535812285" watchObservedRunningTime="2025-12-03 09:42:17.967753055 +0000 UTC m=+3858.536132314" Dec 03 09:42:22 crc kubenswrapper[4573]: I1203 09:42:22.031612 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:42:22 crc kubenswrapper[4573]: E1203 09:42:22.032287 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:42:22 crc kubenswrapper[4573]: I1203 09:42:22.535606 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:22 crc kubenswrapper[4573]: I1203 09:42:22.535673 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:22 crc kubenswrapper[4573]: I1203 09:42:22.587552 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:23 crc kubenswrapper[4573]: I1203 09:42:23.048039 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:23 crc kubenswrapper[4573]: I1203 09:42:23.116126 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6fg8l"] Dec 03 09:42:24 crc kubenswrapper[4573]: I1203 09:42:24.998291 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6fg8l" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerName="registry-server" containerID="cri-o://b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b" gracePeriod=2 Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.621871 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.726520 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g76qn\" (UniqueName: \"kubernetes.io/projected/b529dd20-dbe0-4643-a289-8a3c9f882431-kube-api-access-g76qn\") pod \"b529dd20-dbe0-4643-a289-8a3c9f882431\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.726580 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-utilities\") pod \"b529dd20-dbe0-4643-a289-8a3c9f882431\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.726971 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-catalog-content\") pod \"b529dd20-dbe0-4643-a289-8a3c9f882431\" (UID: \"b529dd20-dbe0-4643-a289-8a3c9f882431\") " Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.727960 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-utilities" (OuterVolumeSpecName: "utilities") pod "b529dd20-dbe0-4643-a289-8a3c9f882431" (UID: "b529dd20-dbe0-4643-a289-8a3c9f882431"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.743694 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b529dd20-dbe0-4643-a289-8a3c9f882431-kube-api-access-g76qn" (OuterVolumeSpecName: "kube-api-access-g76qn") pod "b529dd20-dbe0-4643-a289-8a3c9f882431" (UID: "b529dd20-dbe0-4643-a289-8a3c9f882431"). InnerVolumeSpecName "kube-api-access-g76qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.792564 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b529dd20-dbe0-4643-a289-8a3c9f882431" (UID: "b529dd20-dbe0-4643-a289-8a3c9f882431"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.829545 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g76qn\" (UniqueName: \"kubernetes.io/projected/b529dd20-dbe0-4643-a289-8a3c9f882431-kube-api-access-g76qn\") on node \"crc\" DevicePath \"\"" Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.829591 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:42:25 crc kubenswrapper[4573]: I1203 09:42:25.829601 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b529dd20-dbe0-4643-a289-8a3c9f882431-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.011916 4573 generic.go:334] "Generic (PLEG): container finished" podID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerID="b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b" exitCode=0 Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.012016 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fg8l" event={"ID":"b529dd20-dbe0-4643-a289-8a3c9f882431","Type":"ContainerDied","Data":"b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b"} Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.013292 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6fg8l" event={"ID":"b529dd20-dbe0-4643-a289-8a3c9f882431","Type":"ContainerDied","Data":"c413768d0967208019897a51f7aec031563057daef2af7be5be1d158632dd6b8"} Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.013360 4573 scope.go:117] "RemoveContainer" containerID="b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.012033 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6fg8l" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.072345 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6fg8l"] Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.073680 4573 scope.go:117] "RemoveContainer" containerID="5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.082271 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6fg8l"] Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.110882 4573 scope.go:117] "RemoveContainer" containerID="fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.149627 4573 scope.go:117] "RemoveContainer" containerID="b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b" Dec 03 09:42:26 crc kubenswrapper[4573]: E1203 09:42:26.150136 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b\": container with ID starting with b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b not found: ID does not exist" containerID="b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.150274 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b"} err="failed to get container status \"b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b\": rpc error: code = NotFound desc = could not find container \"b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b\": container with ID starting with b163dcef964c53a274847d2f8defb7139d469a1310d750f0c36b83a227b8793b not found: ID does not exist" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.150315 4573 scope.go:117] "RemoveContainer" containerID="5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1" Dec 03 09:42:26 crc kubenswrapper[4573]: E1203 09:42:26.150685 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1\": container with ID starting with 5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1 not found: ID does not exist" containerID="5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.150724 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1"} err="failed to get container status \"5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1\": rpc error: code = NotFound desc = could not find container \"5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1\": container with ID starting with 5f76ea92c6604eb4983329db0bad773afd274ba8023c1edf488ac58ddb494fe1 not found: ID does not exist" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.150747 4573 scope.go:117] "RemoveContainer" containerID="fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f" Dec 03 09:42:26 crc kubenswrapper[4573]: E1203 09:42:26.151380 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f\": container with ID starting with fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f not found: ID does not exist" containerID="fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f" Dec 03 09:42:26 crc kubenswrapper[4573]: I1203 09:42:26.151425 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f"} err="failed to get container status \"fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f\": rpc error: code = NotFound desc = could not find container \"fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f\": container with ID starting with fbd8405edc7d6d2e4570b268ad5830e504ff2e6af502847787817b16289b0d8f not found: ID does not exist" Dec 03 09:42:28 crc kubenswrapper[4573]: I1203 09:42:28.044884 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" path="/var/lib/kubelet/pods/b529dd20-dbe0-4643-a289-8a3c9f882431/volumes" Dec 03 09:42:35 crc kubenswrapper[4573]: I1203 09:42:35.031676 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:42:35 crc kubenswrapper[4573]: E1203 09:42:35.032802 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:42:48 crc kubenswrapper[4573]: I1203 09:42:48.030150 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:42:48 crc kubenswrapper[4573]: E1203 09:42:48.031908 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:43:02 crc kubenswrapper[4573]: I1203 09:43:02.032165 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:43:02 crc kubenswrapper[4573]: E1203 09:43:02.033010 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:43:15 crc kubenswrapper[4573]: I1203 09:43:15.031291 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:43:15 crc kubenswrapper[4573]: E1203 09:43:15.032696 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:43:27 crc kubenswrapper[4573]: I1203 09:43:27.030657 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:43:27 crc kubenswrapper[4573]: E1203 09:43:27.031475 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:43:42 crc kubenswrapper[4573]: I1203 09:43:42.031333 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:43:42 crc kubenswrapper[4573]: E1203 09:43:42.033062 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:43:55 crc kubenswrapper[4573]: I1203 09:43:55.031095 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:43:55 crc kubenswrapper[4573]: E1203 09:43:55.031871 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:44:07 crc kubenswrapper[4573]: I1203 09:44:07.032013 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:44:07 crc kubenswrapper[4573]: E1203 09:44:07.032898 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:44:22 crc kubenswrapper[4573]: I1203 09:44:22.030798 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:44:22 crc kubenswrapper[4573]: E1203 09:44:22.031585 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:44:37 crc kubenswrapper[4573]: I1203 09:44:37.031547 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:44:37 crc kubenswrapper[4573]: E1203 09:44:37.032278 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:44:48 crc kubenswrapper[4573]: I1203 09:44:48.030782 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:44:48 crc kubenswrapper[4573]: E1203 09:44:48.032287 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.195235 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99"] Dec 03 09:45:00 crc kubenswrapper[4573]: E1203 09:45:00.196306 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerName="extract-utilities" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.196323 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerName="extract-utilities" Dec 03 09:45:00 crc kubenswrapper[4573]: E1203 09:45:00.196351 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerName="extract-content" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.196361 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerName="extract-content" Dec 03 09:45:00 crc kubenswrapper[4573]: E1203 09:45:00.196385 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerName="registry-server" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.196392 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerName="registry-server" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.196566 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="b529dd20-dbe0-4643-a289-8a3c9f882431" containerName="registry-server" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.197192 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.200160 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.214421 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.216315 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99"] Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.290924 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf991cd7-e3da-4100-85be-6057635390ce-config-volume\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.291003 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pjlb\" (UniqueName: \"kubernetes.io/projected/cf991cd7-e3da-4100-85be-6057635390ce-kube-api-access-5pjlb\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.291082 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf991cd7-e3da-4100-85be-6057635390ce-secret-volume\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.394172 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf991cd7-e3da-4100-85be-6057635390ce-config-volume\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.394239 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pjlb\" (UniqueName: \"kubernetes.io/projected/cf991cd7-e3da-4100-85be-6057635390ce-kube-api-access-5pjlb\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.394298 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf991cd7-e3da-4100-85be-6057635390ce-secret-volume\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.395177 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf991cd7-e3da-4100-85be-6057635390ce-config-volume\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.408895 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf991cd7-e3da-4100-85be-6057635390ce-secret-volume\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.414993 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pjlb\" (UniqueName: \"kubernetes.io/projected/cf991cd7-e3da-4100-85be-6057635390ce-kube-api-access-5pjlb\") pod \"collect-profiles-29412585-78d99\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:00 crc kubenswrapper[4573]: I1203 09:45:00.520635 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:01 crc kubenswrapper[4573]: I1203 09:45:01.031872 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:45:01 crc kubenswrapper[4573]: E1203 09:45:01.032644 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:45:01 crc kubenswrapper[4573]: I1203 09:45:01.058574 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99"] Dec 03 09:45:01 crc kubenswrapper[4573]: I1203 09:45:01.590759 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" event={"ID":"cf991cd7-e3da-4100-85be-6057635390ce","Type":"ContainerStarted","Data":"8bd6be8c437b56def23943456d5350397cbcaa518478b208a0db81e9a68f65ba"} Dec 03 09:45:01 crc kubenswrapper[4573]: I1203 09:45:01.591420 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" event={"ID":"cf991cd7-e3da-4100-85be-6057635390ce","Type":"ContainerStarted","Data":"79ee2e299f86e0133686dc816f5df8e8bcc2a8fc660755546f6eecc4556941ab"} Dec 03 09:45:01 crc kubenswrapper[4573]: I1203 09:45:01.613223 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" podStartSLOduration=1.6132053 podStartE2EDuration="1.6132053s" podCreationTimestamp="2025-12-03 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:45:01.607660079 +0000 UTC m=+4022.176039418" watchObservedRunningTime="2025-12-03 09:45:01.6132053 +0000 UTC m=+4022.181584559" Dec 03 09:45:02 crc kubenswrapper[4573]: I1203 09:45:02.605706 4573 generic.go:334] "Generic (PLEG): container finished" podID="cf991cd7-e3da-4100-85be-6057635390ce" containerID="8bd6be8c437b56def23943456d5350397cbcaa518478b208a0db81e9a68f65ba" exitCode=0 Dec 03 09:45:02 crc kubenswrapper[4573]: I1203 09:45:02.605768 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" event={"ID":"cf991cd7-e3da-4100-85be-6057635390ce","Type":"ContainerDied","Data":"8bd6be8c437b56def23943456d5350397cbcaa518478b208a0db81e9a68f65ba"} Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.535877 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.596892 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf991cd7-e3da-4100-85be-6057635390ce-config-volume\") pod \"cf991cd7-e3da-4100-85be-6057635390ce\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.597226 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf991cd7-e3da-4100-85be-6057635390ce-secret-volume\") pod \"cf991cd7-e3da-4100-85be-6057635390ce\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.597255 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pjlb\" (UniqueName: \"kubernetes.io/projected/cf991cd7-e3da-4100-85be-6057635390ce-kube-api-access-5pjlb\") pod \"cf991cd7-e3da-4100-85be-6057635390ce\" (UID: \"cf991cd7-e3da-4100-85be-6057635390ce\") " Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.597580 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf991cd7-e3da-4100-85be-6057635390ce-config-volume" (OuterVolumeSpecName: "config-volume") pod "cf991cd7-e3da-4100-85be-6057635390ce" (UID: "cf991cd7-e3da-4100-85be-6057635390ce"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.598794 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf991cd7-e3da-4100-85be-6057635390ce-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.603455 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf991cd7-e3da-4100-85be-6057635390ce-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cf991cd7-e3da-4100-85be-6057635390ce" (UID: "cf991cd7-e3da-4100-85be-6057635390ce"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.626659 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf991cd7-e3da-4100-85be-6057635390ce-kube-api-access-5pjlb" (OuterVolumeSpecName: "kube-api-access-5pjlb") pod "cf991cd7-e3da-4100-85be-6057635390ce" (UID: "cf991cd7-e3da-4100-85be-6057635390ce"). InnerVolumeSpecName "kube-api-access-5pjlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.634482 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" event={"ID":"cf991cd7-e3da-4100-85be-6057635390ce","Type":"ContainerDied","Data":"79ee2e299f86e0133686dc816f5df8e8bcc2a8fc660755546f6eecc4556941ab"} Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.634523 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79ee2e299f86e0133686dc816f5df8e8bcc2a8fc660755546f6eecc4556941ab" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.634524 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412585-78d99" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.700680 4573 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf991cd7-e3da-4100-85be-6057635390ce-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.700712 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pjlb\" (UniqueName: \"kubernetes.io/projected/cf991cd7-e3da-4100-85be-6057635390ce-kube-api-access-5pjlb\") on node \"crc\" DevicePath \"\"" Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.710352 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv"] Dec 03 09:45:04 crc kubenswrapper[4573]: I1203 09:45:04.718521 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412540-2ddxv"] Dec 03 09:45:06 crc kubenswrapper[4573]: I1203 09:45:06.066631 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561700fb-a6d4-42e1-b339-f3dd062a504a" path="/var/lib/kubelet/pods/561700fb-a6d4-42e1-b339-f3dd062a504a/volumes" Dec 03 09:45:09 crc kubenswrapper[4573]: I1203 09:45:09.130961 4573 scope.go:117] "RemoveContainer" containerID="97bc403af90b64013e601cd9d66868509687cd01e94210a046eb56bdc4d9e1fa" Dec 03 09:45:15 crc kubenswrapper[4573]: I1203 09:45:15.208014 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:45:15 crc kubenswrapper[4573]: E1203 09:45:15.208818 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:45:30 crc kubenswrapper[4573]: I1203 09:45:30.056571 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:45:30 crc kubenswrapper[4573]: E1203 09:45:30.058188 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:45:42 crc kubenswrapper[4573]: I1203 09:45:42.030478 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:45:42 crc kubenswrapper[4573]: E1203 09:45:42.031186 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:45:57 crc kubenswrapper[4573]: I1203 09:45:57.030427 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:45:57 crc kubenswrapper[4573]: E1203 09:45:57.031288 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:46:08 crc kubenswrapper[4573]: I1203 09:46:08.030301 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:46:08 crc kubenswrapper[4573]: E1203 09:46:08.031088 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:46:19 crc kubenswrapper[4573]: I1203 09:46:19.031109 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:46:19 crc kubenswrapper[4573]: E1203 09:46:19.032208 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:46:31 crc kubenswrapper[4573]: I1203 09:46:31.031475 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:46:31 crc kubenswrapper[4573]: E1203 09:46:31.032631 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:46:43 crc kubenswrapper[4573]: I1203 09:46:43.031191 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:46:43 crc kubenswrapper[4573]: E1203 09:46:43.032034 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:46:55 crc kubenswrapper[4573]: I1203 09:46:55.030621 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:46:55 crc kubenswrapper[4573]: E1203 09:46:55.031372 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:47:06 crc kubenswrapper[4573]: I1203 09:47:06.034628 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:47:06 crc kubenswrapper[4573]: I1203 09:47:06.935989 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"46c49ecc556f579a3da2201ded7743f2b47402e58a0209723673b22a9bbbe7f4"} Dec 03 09:47:08 crc kubenswrapper[4573]: I1203 09:47:08.999436 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jttp2"] Dec 03 09:47:09 crc kubenswrapper[4573]: E1203 09:47:09.000375 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf991cd7-e3da-4100-85be-6057635390ce" containerName="collect-profiles" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.000403 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf991cd7-e3da-4100-85be-6057635390ce" containerName="collect-profiles" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.000609 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf991cd7-e3da-4100-85be-6057635390ce" containerName="collect-profiles" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.002373 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.022526 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jttp2"] Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.182685 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd548d8-d599-4da1-b815-287a6a4abdab-utilities\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.182836 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28pwj\" (UniqueName: \"kubernetes.io/projected/cdd548d8-d599-4da1-b815-287a6a4abdab-kube-api-access-28pwj\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.182864 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd548d8-d599-4da1-b815-287a6a4abdab-catalog-content\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.284867 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd548d8-d599-4da1-b815-287a6a4abdab-utilities\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.285016 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28pwj\" (UniqueName: \"kubernetes.io/projected/cdd548d8-d599-4da1-b815-287a6a4abdab-kube-api-access-28pwj\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.285060 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd548d8-d599-4da1-b815-287a6a4abdab-catalog-content\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.285426 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdd548d8-d599-4da1-b815-287a6a4abdab-utilities\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.285517 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdd548d8-d599-4da1-b815-287a6a4abdab-catalog-content\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.311118 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28pwj\" (UniqueName: \"kubernetes.io/projected/cdd548d8-d599-4da1-b815-287a6a4abdab-kube-api-access-28pwj\") pod \"certified-operators-jttp2\" (UID: \"cdd548d8-d599-4da1-b815-287a6a4abdab\") " pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:09 crc kubenswrapper[4573]: I1203 09:47:09.332028 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:10 crc kubenswrapper[4573]: I1203 09:47:10.089019 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jttp2"] Dec 03 09:47:10 crc kubenswrapper[4573]: I1203 09:47:10.985200 4573 generic.go:334] "Generic (PLEG): container finished" podID="cdd548d8-d599-4da1-b815-287a6a4abdab" containerID="49dd34a440885078caa7e4692e2443b5e584d8103fbbbd7bda787f2ebd49c1ab" exitCode=0 Dec 03 09:47:10 crc kubenswrapper[4573]: I1203 09:47:10.985602 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jttp2" event={"ID":"cdd548d8-d599-4da1-b815-287a6a4abdab","Type":"ContainerDied","Data":"49dd34a440885078caa7e4692e2443b5e584d8103fbbbd7bda787f2ebd49c1ab"} Dec 03 09:47:10 crc kubenswrapper[4573]: I1203 09:47:10.985629 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jttp2" event={"ID":"cdd548d8-d599-4da1-b815-287a6a4abdab","Type":"ContainerStarted","Data":"2301e6069229f4e05dd846f638f48f3b7ecd55413887709c8d76e210628a45e6"} Dec 03 09:47:10 crc kubenswrapper[4573]: I1203 09:47:10.988296 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:47:20 crc kubenswrapper[4573]: I1203 09:47:20.083441 4573 generic.go:334] "Generic (PLEG): container finished" podID="cdd548d8-d599-4da1-b815-287a6a4abdab" containerID="4be4cafb89198861dcf47b50c77a8437535d33613f684f9b31072268de1f13e1" exitCode=0 Dec 03 09:47:20 crc kubenswrapper[4573]: I1203 09:47:20.083494 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jttp2" event={"ID":"cdd548d8-d599-4da1-b815-287a6a4abdab","Type":"ContainerDied","Data":"4be4cafb89198861dcf47b50c77a8437535d33613f684f9b31072268de1f13e1"} Dec 03 09:47:22 crc kubenswrapper[4573]: I1203 09:47:22.151718 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jttp2" event={"ID":"cdd548d8-d599-4da1-b815-287a6a4abdab","Type":"ContainerStarted","Data":"8dfe4d9cf1ce441f30771d735b679db2616d36e5f2e345d29d275bd831d4afc3"} Dec 03 09:47:22 crc kubenswrapper[4573]: I1203 09:47:22.220333 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jttp2" podStartSLOduration=3.4758384270000002 podStartE2EDuration="14.220286705s" podCreationTimestamp="2025-12-03 09:47:08 +0000 UTC" firstStartedPulling="2025-12-03 09:47:10.98788531 +0000 UTC m=+4151.556264569" lastFinishedPulling="2025-12-03 09:47:21.732333588 +0000 UTC m=+4162.300712847" observedRunningTime="2025-12-03 09:47:22.19113284 +0000 UTC m=+4162.759512119" watchObservedRunningTime="2025-12-03 09:47:22.220286705 +0000 UTC m=+4162.788665964" Dec 03 09:47:29 crc kubenswrapper[4573]: I1203 09:47:29.334060 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:29 crc kubenswrapper[4573]: I1203 09:47:29.334618 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:29 crc kubenswrapper[4573]: I1203 09:47:29.686321 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:30 crc kubenswrapper[4573]: I1203 09:47:30.284409 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jttp2" Dec 03 09:47:30 crc kubenswrapper[4573]: I1203 09:47:30.362983 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jttp2"] Dec 03 09:47:30 crc kubenswrapper[4573]: I1203 09:47:30.422882 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6zq2f"] Dec 03 09:47:30 crc kubenswrapper[4573]: I1203 09:47:30.423326 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6zq2f" podUID="650903df-cc41-472a-86e1-9c890b4a135c" containerName="registry-server" containerID="cri-o://3b2ee87793e9ed98386658a8263f893e4713acb7d84c761c68b146b9ea636b25" gracePeriod=2 Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.252311 4573 generic.go:334] "Generic (PLEG): container finished" podID="650903df-cc41-472a-86e1-9c890b4a135c" containerID="3b2ee87793e9ed98386658a8263f893e4713acb7d84c761c68b146b9ea636b25" exitCode=0 Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.253169 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zq2f" event={"ID":"650903df-cc41-472a-86e1-9c890b4a135c","Type":"ContainerDied","Data":"3b2ee87793e9ed98386658a8263f893e4713acb7d84c761c68b146b9ea636b25"} Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.713773 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.793547 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncrmj\" (UniqueName: \"kubernetes.io/projected/650903df-cc41-472a-86e1-9c890b4a135c-kube-api-access-ncrmj\") pod \"650903df-cc41-472a-86e1-9c890b4a135c\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.793660 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-catalog-content\") pod \"650903df-cc41-472a-86e1-9c890b4a135c\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.793745 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-utilities\") pod \"650903df-cc41-472a-86e1-9c890b4a135c\" (UID: \"650903df-cc41-472a-86e1-9c890b4a135c\") " Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.795042 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-utilities" (OuterVolumeSpecName: "utilities") pod "650903df-cc41-472a-86e1-9c890b4a135c" (UID: "650903df-cc41-472a-86e1-9c890b4a135c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.807013 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/650903df-cc41-472a-86e1-9c890b4a135c-kube-api-access-ncrmj" (OuterVolumeSpecName: "kube-api-access-ncrmj") pod "650903df-cc41-472a-86e1-9c890b4a135c" (UID: "650903df-cc41-472a-86e1-9c890b4a135c"). InnerVolumeSpecName "kube-api-access-ncrmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.895908 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncrmj\" (UniqueName: \"kubernetes.io/projected/650903df-cc41-472a-86e1-9c890b4a135c-kube-api-access-ncrmj\") on node \"crc\" DevicePath \"\"" Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.895939 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.902887 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "650903df-cc41-472a-86e1-9c890b4a135c" (UID: "650903df-cc41-472a-86e1-9c890b4a135c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:47:31 crc kubenswrapper[4573]: I1203 09:47:31.997855 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/650903df-cc41-472a-86e1-9c890b4a135c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:47:32 crc kubenswrapper[4573]: I1203 09:47:32.263727 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6zq2f" Dec 03 09:47:32 crc kubenswrapper[4573]: I1203 09:47:32.264429 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6zq2f" event={"ID":"650903df-cc41-472a-86e1-9c890b4a135c","Type":"ContainerDied","Data":"298f1ff2b071f9b11c13528d91dc59b4bfd69cc44227cf7e48a1dc147d4cf501"} Dec 03 09:47:32 crc kubenswrapper[4573]: I1203 09:47:32.264488 4573 scope.go:117] "RemoveContainer" containerID="3b2ee87793e9ed98386658a8263f893e4713acb7d84c761c68b146b9ea636b25" Dec 03 09:47:32 crc kubenswrapper[4573]: I1203 09:47:32.300186 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6zq2f"] Dec 03 09:47:32 crc kubenswrapper[4573]: I1203 09:47:32.308722 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6zq2f"] Dec 03 09:47:32 crc kubenswrapper[4573]: I1203 09:47:32.335946 4573 scope.go:117] "RemoveContainer" containerID="80ff6a52d9927751b873cc37b8fe346eaec45ce6978fa671044aef9fe6bd928a" Dec 03 09:47:32 crc kubenswrapper[4573]: I1203 09:47:32.438828 4573 scope.go:117] "RemoveContainer" containerID="8e1d4e2a61b2d6a46ef5ebab4e50b91365768270f1547541cdad5e5bd7503afd" Dec 03 09:47:34 crc kubenswrapper[4573]: I1203 09:47:34.052257 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="650903df-cc41-472a-86e1-9c890b4a135c" path="/var/lib/kubelet/pods/650903df-cc41-472a-86e1-9c890b4a135c/volumes" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.107228 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6tnlf"] Dec 03 09:47:45 crc kubenswrapper[4573]: E1203 09:47:45.108140 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650903df-cc41-472a-86e1-9c890b4a135c" containerName="extract-content" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.108153 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="650903df-cc41-472a-86e1-9c890b4a135c" containerName="extract-content" Dec 03 09:47:45 crc kubenswrapper[4573]: E1203 09:47:45.108166 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650903df-cc41-472a-86e1-9c890b4a135c" containerName="registry-server" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.108172 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="650903df-cc41-472a-86e1-9c890b4a135c" containerName="registry-server" Dec 03 09:47:45 crc kubenswrapper[4573]: E1203 09:47:45.108190 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650903df-cc41-472a-86e1-9c890b4a135c" containerName="extract-utilities" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.108217 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="650903df-cc41-472a-86e1-9c890b4a135c" containerName="extract-utilities" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.108446 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="650903df-cc41-472a-86e1-9c890b4a135c" containerName="registry-server" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.110703 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.124916 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tnlf"] Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.186396 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-utilities\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.186864 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x99gm\" (UniqueName: \"kubernetes.io/projected/cee90265-9423-4f4e-8d6a-0b1886e5dc30-kube-api-access-x99gm\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.187074 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-catalog-content\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.289505 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-utilities\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.289653 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x99gm\" (UniqueName: \"kubernetes.io/projected/cee90265-9423-4f4e-8d6a-0b1886e5dc30-kube-api-access-x99gm\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.289688 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-catalog-content\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.289977 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-utilities\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.290498 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-catalog-content\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.312507 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x99gm\" (UniqueName: \"kubernetes.io/projected/cee90265-9423-4f4e-8d6a-0b1886e5dc30-kube-api-access-x99gm\") pod \"redhat-marketplace-6tnlf\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:45 crc kubenswrapper[4573]: I1203 09:47:45.444011 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:46 crc kubenswrapper[4573]: I1203 09:47:46.050615 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tnlf"] Dec 03 09:47:46 crc kubenswrapper[4573]: I1203 09:47:46.412350 4573 generic.go:334] "Generic (PLEG): container finished" podID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerID="2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe" exitCode=0 Dec 03 09:47:46 crc kubenswrapper[4573]: I1203 09:47:46.412410 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tnlf" event={"ID":"cee90265-9423-4f4e-8d6a-0b1886e5dc30","Type":"ContainerDied","Data":"2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe"} Dec 03 09:47:46 crc kubenswrapper[4573]: I1203 09:47:46.412624 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tnlf" event={"ID":"cee90265-9423-4f4e-8d6a-0b1886e5dc30","Type":"ContainerStarted","Data":"94c99f060ae1dad74f646bfa7cb304f344e3abbb2c45f7e750b9c006f3940501"} Dec 03 09:47:47 crc kubenswrapper[4573]: I1203 09:47:47.425459 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tnlf" event={"ID":"cee90265-9423-4f4e-8d6a-0b1886e5dc30","Type":"ContainerStarted","Data":"498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b"} Dec 03 09:47:48 crc kubenswrapper[4573]: I1203 09:47:48.438335 4573 generic.go:334] "Generic (PLEG): container finished" podID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerID="498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b" exitCode=0 Dec 03 09:47:48 crc kubenswrapper[4573]: I1203 09:47:48.438455 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tnlf" event={"ID":"cee90265-9423-4f4e-8d6a-0b1886e5dc30","Type":"ContainerDied","Data":"498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b"} Dec 03 09:47:49 crc kubenswrapper[4573]: I1203 09:47:49.451357 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tnlf" event={"ID":"cee90265-9423-4f4e-8d6a-0b1886e5dc30","Type":"ContainerStarted","Data":"3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c"} Dec 03 09:47:49 crc kubenswrapper[4573]: I1203 09:47:49.480206 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6tnlf" podStartSLOduration=1.916961124 podStartE2EDuration="4.480187035s" podCreationTimestamp="2025-12-03 09:47:45 +0000 UTC" firstStartedPulling="2025-12-03 09:47:46.415215631 +0000 UTC m=+4186.983594900" lastFinishedPulling="2025-12-03 09:47:48.978441552 +0000 UTC m=+4189.546820811" observedRunningTime="2025-12-03 09:47:49.473947085 +0000 UTC m=+4190.042326344" watchObservedRunningTime="2025-12-03 09:47:49.480187035 +0000 UTC m=+4190.048566294" Dec 03 09:47:55 crc kubenswrapper[4573]: I1203 09:47:55.445175 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:55 crc kubenswrapper[4573]: I1203 09:47:55.446005 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:55 crc kubenswrapper[4573]: I1203 09:47:55.516719 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:55 crc kubenswrapper[4573]: I1203 09:47:55.575323 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:55 crc kubenswrapper[4573]: I1203 09:47:55.780520 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tnlf"] Dec 03 09:47:57 crc kubenswrapper[4573]: I1203 09:47:57.542225 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6tnlf" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerName="registry-server" containerID="cri-o://3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c" gracePeriod=2 Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.330407 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.467428 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x99gm\" (UniqueName: \"kubernetes.io/projected/cee90265-9423-4f4e-8d6a-0b1886e5dc30-kube-api-access-x99gm\") pod \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.467714 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-utilities\") pod \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.467833 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-catalog-content\") pod \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\" (UID: \"cee90265-9423-4f4e-8d6a-0b1886e5dc30\") " Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.468783 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-utilities" (OuterVolumeSpecName: "utilities") pod "cee90265-9423-4f4e-8d6a-0b1886e5dc30" (UID: "cee90265-9423-4f4e-8d6a-0b1886e5dc30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.469408 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.486425 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee90265-9423-4f4e-8d6a-0b1886e5dc30-kube-api-access-x99gm" (OuterVolumeSpecName: "kube-api-access-x99gm") pod "cee90265-9423-4f4e-8d6a-0b1886e5dc30" (UID: "cee90265-9423-4f4e-8d6a-0b1886e5dc30"). InnerVolumeSpecName "kube-api-access-x99gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.497722 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cee90265-9423-4f4e-8d6a-0b1886e5dc30" (UID: "cee90265-9423-4f4e-8d6a-0b1886e5dc30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.569502 4573 generic.go:334] "Generic (PLEG): container finished" podID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerID="3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c" exitCode=0 Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.569576 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tnlf" event={"ID":"cee90265-9423-4f4e-8d6a-0b1886e5dc30","Type":"ContainerDied","Data":"3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c"} Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.569622 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tnlf" event={"ID":"cee90265-9423-4f4e-8d6a-0b1886e5dc30","Type":"ContainerDied","Data":"94c99f060ae1dad74f646bfa7cb304f344e3abbb2c45f7e750b9c006f3940501"} Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.569655 4573 scope.go:117] "RemoveContainer" containerID="3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.569933 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tnlf" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.574253 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x99gm\" (UniqueName: \"kubernetes.io/projected/cee90265-9423-4f4e-8d6a-0b1886e5dc30-kube-api-access-x99gm\") on node \"crc\" DevicePath \"\"" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.574308 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cee90265-9423-4f4e-8d6a-0b1886e5dc30-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.607296 4573 scope.go:117] "RemoveContainer" containerID="498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.634607 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tnlf"] Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.644230 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tnlf"] Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.663477 4573 scope.go:117] "RemoveContainer" containerID="2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.708427 4573 scope.go:117] "RemoveContainer" containerID="3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c" Dec 03 09:47:58 crc kubenswrapper[4573]: E1203 09:47:58.709938 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c\": container with ID starting with 3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c not found: ID does not exist" containerID="3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.710019 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c"} err="failed to get container status \"3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c\": rpc error: code = NotFound desc = could not find container \"3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c\": container with ID starting with 3f0d4b1cfbc884417f32847c78bac204dc89358767f8f737b040662d472cd55c not found: ID does not exist" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.710083 4573 scope.go:117] "RemoveContainer" containerID="498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b" Dec 03 09:47:58 crc kubenswrapper[4573]: E1203 09:47:58.710596 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b\": container with ID starting with 498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b not found: ID does not exist" containerID="498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.710658 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b"} err="failed to get container status \"498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b\": rpc error: code = NotFound desc = could not find container \"498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b\": container with ID starting with 498d3e6346f92b3af3090e90f87aeaecaac5cee4d77fc5689f2bc39b2f08696b not found: ID does not exist" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.710685 4573 scope.go:117] "RemoveContainer" containerID="2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe" Dec 03 09:47:58 crc kubenswrapper[4573]: E1203 09:47:58.711211 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe\": container with ID starting with 2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe not found: ID does not exist" containerID="2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe" Dec 03 09:47:58 crc kubenswrapper[4573]: I1203 09:47:58.711234 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe"} err="failed to get container status \"2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe\": rpc error: code = NotFound desc = could not find container \"2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe\": container with ID starting with 2c5488d60ccaccea6d8e88331def8e5aba7267f3903c0df527691f50979319fe not found: ID does not exist" Dec 03 09:48:00 crc kubenswrapper[4573]: I1203 09:48:00.068886 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" path="/var/lib/kubelet/pods/cee90265-9423-4f4e-8d6a-0b1886e5dc30/volumes" Dec 03 09:49:26 crc kubenswrapper[4573]: I1203 09:49:26.943698 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:49:26 crc kubenswrapper[4573]: I1203 09:49:26.944291 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.537486 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kvjjs"] Dec 03 09:49:27 crc kubenswrapper[4573]: E1203 09:49:27.537981 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerName="registry-server" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.538006 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerName="registry-server" Dec 03 09:49:27 crc kubenswrapper[4573]: E1203 09:49:27.538035 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerName="extract-content" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.538066 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerName="extract-content" Dec 03 09:49:27 crc kubenswrapper[4573]: E1203 09:49:27.538115 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerName="extract-utilities" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.538125 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerName="extract-utilities" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.538360 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee90265-9423-4f4e-8d6a-0b1886e5dc30" containerName="registry-server" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.540077 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.576174 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kvjjs"] Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.628272 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-utilities\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.628317 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v82kg\" (UniqueName: \"kubernetes.io/projected/f8d48056-283a-4b6f-bd03-cf2de540ddfc-kube-api-access-v82kg\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.628436 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-catalog-content\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.730310 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-catalog-content\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.730422 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-utilities\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.730443 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v82kg\" (UniqueName: \"kubernetes.io/projected/f8d48056-283a-4b6f-bd03-cf2de540ddfc-kube-api-access-v82kg\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.731422 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-catalog-content\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.731649 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-utilities\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.849279 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v82kg\" (UniqueName: \"kubernetes.io/projected/f8d48056-283a-4b6f-bd03-cf2de540ddfc-kube-api-access-v82kg\") pod \"redhat-operators-kvjjs\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:27 crc kubenswrapper[4573]: I1203 09:49:27.863555 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:28 crc kubenswrapper[4573]: I1203 09:49:28.460306 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kvjjs"] Dec 03 09:49:29 crc kubenswrapper[4573]: I1203 09:49:29.551641 4573 generic.go:334] "Generic (PLEG): container finished" podID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerID="eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd" exitCode=0 Dec 03 09:49:29 crc kubenswrapper[4573]: I1203 09:49:29.551838 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvjjs" event={"ID":"f8d48056-283a-4b6f-bd03-cf2de540ddfc","Type":"ContainerDied","Data":"eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd"} Dec 03 09:49:29 crc kubenswrapper[4573]: I1203 09:49:29.552210 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvjjs" event={"ID":"f8d48056-283a-4b6f-bd03-cf2de540ddfc","Type":"ContainerStarted","Data":"a27f4cc1149d2abf2fd2c99f1c27391aa90abe1faca936be2795af976cf43d3c"} Dec 03 09:49:30 crc kubenswrapper[4573]: I1203 09:49:30.574172 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvjjs" event={"ID":"f8d48056-283a-4b6f-bd03-cf2de540ddfc","Type":"ContainerStarted","Data":"908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52"} Dec 03 09:49:34 crc kubenswrapper[4573]: I1203 09:49:34.615554 4573 generic.go:334] "Generic (PLEG): container finished" podID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerID="908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52" exitCode=0 Dec 03 09:49:34 crc kubenswrapper[4573]: I1203 09:49:34.615642 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvjjs" event={"ID":"f8d48056-283a-4b6f-bd03-cf2de540ddfc","Type":"ContainerDied","Data":"908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52"} Dec 03 09:49:35 crc kubenswrapper[4573]: I1203 09:49:35.629070 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvjjs" event={"ID":"f8d48056-283a-4b6f-bd03-cf2de540ddfc","Type":"ContainerStarted","Data":"ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d"} Dec 03 09:49:37 crc kubenswrapper[4573]: I1203 09:49:37.865495 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:37 crc kubenswrapper[4573]: I1203 09:49:37.866742 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:38 crc kubenswrapper[4573]: I1203 09:49:38.909609 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kvjjs" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="registry-server" probeResult="failure" output=< Dec 03 09:49:38 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:49:38 crc kubenswrapper[4573]: > Dec 03 09:49:49 crc kubenswrapper[4573]: I1203 09:49:49.365705 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kvjjs" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="registry-server" probeResult="failure" output=< Dec 03 09:49:49 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:49:49 crc kubenswrapper[4573]: > Dec 03 09:49:56 crc kubenswrapper[4573]: I1203 09:49:56.942984 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:49:56 crc kubenswrapper[4573]: I1203 09:49:56.943551 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:49:57 crc kubenswrapper[4573]: I1203 09:49:57.925119 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:57 crc kubenswrapper[4573]: I1203 09:49:57.949930 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kvjjs" podStartSLOduration=25.513889203 podStartE2EDuration="30.949899279s" podCreationTimestamp="2025-12-03 09:49:27 +0000 UTC" firstStartedPulling="2025-12-03 09:49:29.554160566 +0000 UTC m=+4290.122539825" lastFinishedPulling="2025-12-03 09:49:34.990170642 +0000 UTC m=+4295.558549901" observedRunningTime="2025-12-03 09:49:35.746804261 +0000 UTC m=+4296.315183540" watchObservedRunningTime="2025-12-03 09:49:57.949899279 +0000 UTC m=+4318.518278538" Dec 03 09:49:57 crc kubenswrapper[4573]: I1203 09:49:57.985026 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:49:58 crc kubenswrapper[4573]: I1203 09:49:58.744545 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kvjjs"] Dec 03 09:49:59 crc kubenswrapper[4573]: I1203 09:49:59.917222 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kvjjs" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="registry-server" containerID="cri-o://ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d" gracePeriod=2 Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.829136 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.918533 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-catalog-content\") pod \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.918577 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-utilities\") pod \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.918616 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v82kg\" (UniqueName: \"kubernetes.io/projected/f8d48056-283a-4b6f-bd03-cf2de540ddfc-kube-api-access-v82kg\") pod \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\" (UID: \"f8d48056-283a-4b6f-bd03-cf2de540ddfc\") " Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.919496 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-utilities" (OuterVolumeSpecName: "utilities") pod "f8d48056-283a-4b6f-bd03-cf2de540ddfc" (UID: "f8d48056-283a-4b6f-bd03-cf2de540ddfc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.927568 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d48056-283a-4b6f-bd03-cf2de540ddfc-kube-api-access-v82kg" (OuterVolumeSpecName: "kube-api-access-v82kg") pod "f8d48056-283a-4b6f-bd03-cf2de540ddfc" (UID: "f8d48056-283a-4b6f-bd03-cf2de540ddfc"). InnerVolumeSpecName "kube-api-access-v82kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.931014 4573 generic.go:334] "Generic (PLEG): container finished" podID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerID="ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d" exitCode=0 Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.931074 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvjjs" event={"ID":"f8d48056-283a-4b6f-bd03-cf2de540ddfc","Type":"ContainerDied","Data":"ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d"} Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.931120 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvjjs" Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.931178 4573 scope.go:117] "RemoveContainer" containerID="ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d" Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.931134 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvjjs" event={"ID":"f8d48056-283a-4b6f-bd03-cf2de540ddfc","Type":"ContainerDied","Data":"a27f4cc1149d2abf2fd2c99f1c27391aa90abe1faca936be2795af976cf43d3c"} Dec 03 09:50:00 crc kubenswrapper[4573]: I1203 09:50:00.986974 4573 scope.go:117] "RemoveContainer" containerID="908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.013596 4573 scope.go:117] "RemoveContainer" containerID="eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.021420 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.021476 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v82kg\" (UniqueName: \"kubernetes.io/projected/f8d48056-283a-4b6f-bd03-cf2de540ddfc-kube-api-access-v82kg\") on node \"crc\" DevicePath \"\"" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.039200 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8d48056-283a-4b6f-bd03-cf2de540ddfc" (UID: "f8d48056-283a-4b6f-bd03-cf2de540ddfc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.058654 4573 scope.go:117] "RemoveContainer" containerID="ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d" Dec 03 09:50:01 crc kubenswrapper[4573]: E1203 09:50:01.059525 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d\": container with ID starting with ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d not found: ID does not exist" containerID="ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.059595 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d"} err="failed to get container status \"ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d\": rpc error: code = NotFound desc = could not find container \"ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d\": container with ID starting with ee8a5737722ab384b0f28bb01663e0d52d3c9c9fa3d72e21920832eaf260064d not found: ID does not exist" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.059626 4573 scope.go:117] "RemoveContainer" containerID="908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52" Dec 03 09:50:01 crc kubenswrapper[4573]: E1203 09:50:01.059937 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52\": container with ID starting with 908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52 not found: ID does not exist" containerID="908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.060027 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52"} err="failed to get container status \"908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52\": rpc error: code = NotFound desc = could not find container \"908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52\": container with ID starting with 908bc4ffb06a3ba0e60f7de0f2822b7dc0d7f0cf42633f1c4c683f87d7a06e52 not found: ID does not exist" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.060073 4573 scope.go:117] "RemoveContainer" containerID="eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd" Dec 03 09:50:01 crc kubenswrapper[4573]: E1203 09:50:01.060315 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd\": container with ID starting with eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd not found: ID does not exist" containerID="eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.060340 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd"} err="failed to get container status \"eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd\": rpc error: code = NotFound desc = could not find container \"eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd\": container with ID starting with eb1b7ebb0d914196bcc05d95574cd1a1ff5dad1732f0559e88b42562332a7afd not found: ID does not exist" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.124165 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d48056-283a-4b6f-bd03-cf2de540ddfc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.298210 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kvjjs"] Dec 03 09:50:01 crc kubenswrapper[4573]: I1203 09:50:01.304678 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kvjjs"] Dec 03 09:50:02 crc kubenswrapper[4573]: I1203 09:50:02.042479 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" path="/var/lib/kubelet/pods/f8d48056-283a-4b6f-bd03-cf2de540ddfc/volumes" Dec 03 09:50:26 crc kubenswrapper[4573]: I1203 09:50:26.943087 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:50:26 crc kubenswrapper[4573]: I1203 09:50:26.943712 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:50:26 crc kubenswrapper[4573]: I1203 09:50:26.943767 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:50:26 crc kubenswrapper[4573]: I1203 09:50:26.944730 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46c49ecc556f579a3da2201ded7743f2b47402e58a0209723673b22a9bbbe7f4"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:50:26 crc kubenswrapper[4573]: I1203 09:50:26.944837 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://46c49ecc556f579a3da2201ded7743f2b47402e58a0209723673b22a9bbbe7f4" gracePeriod=600 Dec 03 09:50:27 crc kubenswrapper[4573]: I1203 09:50:27.190626 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="46c49ecc556f579a3da2201ded7743f2b47402e58a0209723673b22a9bbbe7f4" exitCode=0 Dec 03 09:50:27 crc kubenswrapper[4573]: I1203 09:50:27.190687 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"46c49ecc556f579a3da2201ded7743f2b47402e58a0209723673b22a9bbbe7f4"} Dec 03 09:50:27 crc kubenswrapper[4573]: I1203 09:50:27.190859 4573 scope.go:117] "RemoveContainer" containerID="259a3ce318f8243a34f1a39641815499e6c3083d88b9322ca235c7bdd8dbd404" Dec 03 09:50:28 crc kubenswrapper[4573]: I1203 09:50:28.202330 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7"} Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.528769 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dmn65"] Dec 03 09:52:43 crc kubenswrapper[4573]: E1203 09:52:43.529906 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="registry-server" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.529927 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="registry-server" Dec 03 09:52:43 crc kubenswrapper[4573]: E1203 09:52:43.529953 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="extract-content" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.529961 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="extract-content" Dec 03 09:52:43 crc kubenswrapper[4573]: E1203 09:52:43.530011 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="extract-utilities" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.530022 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="extract-utilities" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.530286 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d48056-283a-4b6f-bd03-cf2de540ddfc" containerName="registry-server" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.532461 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.589726 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dmn65"] Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.616781 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-catalog-content\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.616848 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-utilities\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.617036 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vvmd\" (UniqueName: \"kubernetes.io/projected/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-kube-api-access-9vvmd\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.718680 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-catalog-content\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.718736 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-utilities\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.718860 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vvmd\" (UniqueName: \"kubernetes.io/projected/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-kube-api-access-9vvmd\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.719715 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-catalog-content\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.719937 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-utilities\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.743189 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vvmd\" (UniqueName: \"kubernetes.io/projected/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-kube-api-access-9vvmd\") pod \"community-operators-dmn65\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:43 crc kubenswrapper[4573]: I1203 09:52:43.857760 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:44 crc kubenswrapper[4573]: I1203 09:52:44.403698 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dmn65"] Dec 03 09:52:44 crc kubenswrapper[4573]: W1203 09:52:44.413923 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37618308_e7ee_4dc1_9b05_6d2e421f6bfc.slice/crio-0b6b60b241092520c3daa4f93a70305fb4b8507ffa74fa677af44e384baddb0b WatchSource:0}: Error finding container 0b6b60b241092520c3daa4f93a70305fb4b8507ffa74fa677af44e384baddb0b: Status 404 returned error can't find the container with id 0b6b60b241092520c3daa4f93a70305fb4b8507ffa74fa677af44e384baddb0b Dec 03 09:52:44 crc kubenswrapper[4573]: I1203 09:52:44.608359 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dmn65" event={"ID":"37618308-e7ee-4dc1-9b05-6d2e421f6bfc","Type":"ContainerStarted","Data":"0b6b60b241092520c3daa4f93a70305fb4b8507ffa74fa677af44e384baddb0b"} Dec 03 09:52:45 crc kubenswrapper[4573]: I1203 09:52:45.620378 4573 generic.go:334] "Generic (PLEG): container finished" podID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerID="1a3600ed1e3b8aaa3210d35482f1ebe1565ecabd3682a6c106bb3edb1398b771" exitCode=0 Dec 03 09:52:45 crc kubenswrapper[4573]: I1203 09:52:45.620436 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dmn65" event={"ID":"37618308-e7ee-4dc1-9b05-6d2e421f6bfc","Type":"ContainerDied","Data":"1a3600ed1e3b8aaa3210d35482f1ebe1565ecabd3682a6c106bb3edb1398b771"} Dec 03 09:52:45 crc kubenswrapper[4573]: I1203 09:52:45.623692 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:52:47 crc kubenswrapper[4573]: I1203 09:52:47.643968 4573 generic.go:334] "Generic (PLEG): container finished" podID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerID="75ac462573a8c2a31ac28283118ec05991caf5727f87eb48d4dacf9e1f1d7703" exitCode=0 Dec 03 09:52:47 crc kubenswrapper[4573]: I1203 09:52:47.644016 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dmn65" event={"ID":"37618308-e7ee-4dc1-9b05-6d2e421f6bfc","Type":"ContainerDied","Data":"75ac462573a8c2a31ac28283118ec05991caf5727f87eb48d4dacf9e1f1d7703"} Dec 03 09:52:48 crc kubenswrapper[4573]: I1203 09:52:48.657653 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dmn65" event={"ID":"37618308-e7ee-4dc1-9b05-6d2e421f6bfc","Type":"ContainerStarted","Data":"8c9de3f8b9fb847980e0b2a351aa893fd7532f4b87119a738320d2d1b42d9798"} Dec 03 09:52:48 crc kubenswrapper[4573]: I1203 09:52:48.677968 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dmn65" podStartSLOduration=3.180290413 podStartE2EDuration="5.677948077s" podCreationTimestamp="2025-12-03 09:52:43 +0000 UTC" firstStartedPulling="2025-12-03 09:52:45.623346156 +0000 UTC m=+4486.191725415" lastFinishedPulling="2025-12-03 09:52:48.1210038 +0000 UTC m=+4488.689383079" observedRunningTime="2025-12-03 09:52:48.675153331 +0000 UTC m=+4489.243532600" watchObservedRunningTime="2025-12-03 09:52:48.677948077 +0000 UTC m=+4489.246327336" Dec 03 09:52:53 crc kubenswrapper[4573]: I1203 09:52:53.858878 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:53 crc kubenswrapper[4573]: I1203 09:52:53.859464 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:53 crc kubenswrapper[4573]: I1203 09:52:53.904992 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:54 crc kubenswrapper[4573]: I1203 09:52:54.755166 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:54 crc kubenswrapper[4573]: I1203 09:52:54.807121 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dmn65"] Dec 03 09:52:56 crc kubenswrapper[4573]: I1203 09:52:56.721818 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dmn65" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerName="registry-server" containerID="cri-o://8c9de3f8b9fb847980e0b2a351aa893fd7532f4b87119a738320d2d1b42d9798" gracePeriod=2 Dec 03 09:52:56 crc kubenswrapper[4573]: I1203 09:52:56.943132 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:52:56 crc kubenswrapper[4573]: I1203 09:52:56.943212 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:52:57 crc kubenswrapper[4573]: I1203 09:52:57.741250 4573 generic.go:334] "Generic (PLEG): container finished" podID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerID="8c9de3f8b9fb847980e0b2a351aa893fd7532f4b87119a738320d2d1b42d9798" exitCode=0 Dec 03 09:52:57 crc kubenswrapper[4573]: I1203 09:52:57.741545 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dmn65" event={"ID":"37618308-e7ee-4dc1-9b05-6d2e421f6bfc","Type":"ContainerDied","Data":"8c9de3f8b9fb847980e0b2a351aa893fd7532f4b87119a738320d2d1b42d9798"} Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.071436 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.212692 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-catalog-content\") pod \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.212825 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-utilities\") pod \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.212912 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vvmd\" (UniqueName: \"kubernetes.io/projected/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-kube-api-access-9vvmd\") pod \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\" (UID: \"37618308-e7ee-4dc1-9b05-6d2e421f6bfc\") " Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.214008 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-utilities" (OuterVolumeSpecName: "utilities") pod "37618308-e7ee-4dc1-9b05-6d2e421f6bfc" (UID: "37618308-e7ee-4dc1-9b05-6d2e421f6bfc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.220472 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-kube-api-access-9vvmd" (OuterVolumeSpecName: "kube-api-access-9vvmd") pod "37618308-e7ee-4dc1-9b05-6d2e421f6bfc" (UID: "37618308-e7ee-4dc1-9b05-6d2e421f6bfc"). InnerVolumeSpecName "kube-api-access-9vvmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.271299 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37618308-e7ee-4dc1-9b05-6d2e421f6bfc" (UID: "37618308-e7ee-4dc1-9b05-6d2e421f6bfc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.315469 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.315519 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.315534 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vvmd\" (UniqueName: \"kubernetes.io/projected/37618308-e7ee-4dc1-9b05-6d2e421f6bfc-kube-api-access-9vvmd\") on node \"crc\" DevicePath \"\"" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.754722 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dmn65" event={"ID":"37618308-e7ee-4dc1-9b05-6d2e421f6bfc","Type":"ContainerDied","Data":"0b6b60b241092520c3daa4f93a70305fb4b8507ffa74fa677af44e384baddb0b"} Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.754792 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dmn65" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.754818 4573 scope.go:117] "RemoveContainer" containerID="8c9de3f8b9fb847980e0b2a351aa893fd7532f4b87119a738320d2d1b42d9798" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.793490 4573 scope.go:117] "RemoveContainer" containerID="75ac462573a8c2a31ac28283118ec05991caf5727f87eb48d4dacf9e1f1d7703" Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.796489 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dmn65"] Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.807793 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dmn65"] Dec 03 09:52:58 crc kubenswrapper[4573]: I1203 09:52:58.823623 4573 scope.go:117] "RemoveContainer" containerID="1a3600ed1e3b8aaa3210d35482f1ebe1565ecabd3682a6c106bb3edb1398b771" Dec 03 09:53:00 crc kubenswrapper[4573]: I1203 09:53:00.057326 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" path="/var/lib/kubelet/pods/37618308-e7ee-4dc1-9b05-6d2e421f6bfc/volumes" Dec 03 09:53:26 crc kubenswrapper[4573]: I1203 09:53:26.943078 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:53:26 crc kubenswrapper[4573]: I1203 09:53:26.943689 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:53:56 crc kubenswrapper[4573]: I1203 09:53:56.943555 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 09:53:56 crc kubenswrapper[4573]: I1203 09:53:56.944451 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 09:53:56 crc kubenswrapper[4573]: I1203 09:53:56.944505 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 09:53:56 crc kubenswrapper[4573]: I1203 09:53:56.945108 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 09:53:56 crc kubenswrapper[4573]: I1203 09:53:56.945168 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" gracePeriod=600 Dec 03 09:53:57 crc kubenswrapper[4573]: E1203 09:53:57.074416 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:53:57 crc kubenswrapper[4573]: I1203 09:53:57.521330 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" exitCode=0 Dec 03 09:53:57 crc kubenswrapper[4573]: I1203 09:53:57.521383 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7"} Dec 03 09:53:57 crc kubenswrapper[4573]: I1203 09:53:57.521439 4573 scope.go:117] "RemoveContainer" containerID="46c49ecc556f579a3da2201ded7743f2b47402e58a0209723673b22a9bbbe7f4" Dec 03 09:53:57 crc kubenswrapper[4573]: I1203 09:53:57.522210 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:53:57 crc kubenswrapper[4573]: E1203 09:53:57.523132 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:54:12 crc kubenswrapper[4573]: I1203 09:54:12.035162 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:54:12 crc kubenswrapper[4573]: E1203 09:54:12.035864 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:54:23 crc kubenswrapper[4573]: I1203 09:54:23.030970 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:54:23 crc kubenswrapper[4573]: E1203 09:54:23.031822 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:54:34 crc kubenswrapper[4573]: I1203 09:54:34.030575 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:54:34 crc kubenswrapper[4573]: E1203 09:54:34.032456 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:54:45 crc kubenswrapper[4573]: I1203 09:54:45.031515 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:54:45 crc kubenswrapper[4573]: E1203 09:54:45.032576 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:54:57 crc kubenswrapper[4573]: I1203 09:54:57.031909 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:54:57 crc kubenswrapper[4573]: E1203 09:54:57.033145 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:55:08 crc kubenswrapper[4573]: I1203 09:55:08.030565 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:55:08 crc kubenswrapper[4573]: E1203 09:55:08.032764 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:55:22 crc kubenswrapper[4573]: I1203 09:55:22.031737 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:55:22 crc kubenswrapper[4573]: E1203 09:55:22.033325 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:55:37 crc kubenswrapper[4573]: I1203 09:55:37.031269 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:55:37 crc kubenswrapper[4573]: E1203 09:55:37.032169 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:55:49 crc kubenswrapper[4573]: I1203 09:55:49.031575 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:55:49 crc kubenswrapper[4573]: E1203 09:55:49.032533 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:56:00 crc kubenswrapper[4573]: I1203 09:56:00.040846 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:56:00 crc kubenswrapper[4573]: E1203 09:56:00.041813 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:56:02 crc kubenswrapper[4573]: I1203 09:56:02.430534 4573 generic.go:334] "Generic (PLEG): container finished" podID="9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" containerID="f0d81a02756587a09e72e1565d897fa9b12547cbeb84e52b823b02f16aee49c6" exitCode=0 Dec 03 09:56:02 crc kubenswrapper[4573]: I1203 09:56:02.430657 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b","Type":"ContainerDied","Data":"f0d81a02756587a09e72e1565d897fa9b12547cbeb84e52b823b02f16aee49c6"} Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.017683 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.153776 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.153878 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config-secret\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.153931 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87llw\" (UniqueName: \"kubernetes.io/projected/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-kube-api-access-87llw\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.153955 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ssh-key\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.153999 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ca-certs\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.154033 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-temporary\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.154124 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-config-data\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.154159 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-workdir\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.154221 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\" (UID: \"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b\") " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.155483 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.156563 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-config-data" (OuterVolumeSpecName: "config-data") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.161012 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-kube-api-access-87llw" (OuterVolumeSpecName: "kube-api-access-87llw") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "kube-api-access-87llw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.164498 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.185272 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.190234 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.192199 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.208216 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.230283 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" (UID: "9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256032 4573 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256085 4573 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256097 4573 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256107 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87llw\" (UniqueName: \"kubernetes.io/projected/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-kube-api-access-87llw\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256116 4573 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256124 4573 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256133 4573 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256141 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.256151 4573 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.286202 4573 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.357473 4573 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.448164 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b","Type":"ContainerDied","Data":"74a544f0bb525fb50578f612a043af1f5dc0ebed8522ce58ed8adc1f5b86729f"} Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.448226 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74a544f0bb525fb50578f612a043af1f5dc0ebed8522ce58ed8adc1f5b86729f" Dec 03 09:56:04 crc kubenswrapper[4573]: I1203 09:56:04.448297 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.946963 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 09:56:09 crc kubenswrapper[4573]: E1203 09:56:09.947980 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerName="extract-content" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.947995 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerName="extract-content" Dec 03 09:56:09 crc kubenswrapper[4573]: E1203 09:56:09.948020 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerName="extract-utilities" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.948027 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerName="extract-utilities" Dec 03 09:56:09 crc kubenswrapper[4573]: E1203 09:56:09.948073 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" containerName="tempest-tests-tempest-tests-runner" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.948082 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" containerName="tempest-tests-tempest-tests-runner" Dec 03 09:56:09 crc kubenswrapper[4573]: E1203 09:56:09.948096 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerName="registry-server" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.948103 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerName="registry-server" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.948335 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b" containerName="tempest-tests-tempest-tests-runner" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.948351 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="37618308-e7ee-4dc1-9b05-6d2e421f6bfc" containerName="registry-server" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.949255 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.952156 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pfxr8" Dec 03 09:56:09 crc kubenswrapper[4573]: I1203 09:56:09.956348 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.082202 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f5b253a6-9fcb-432b-81d9-6a6b980367c7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.082556 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fbx8\" (UniqueName: \"kubernetes.io/projected/f5b253a6-9fcb-432b-81d9-6a6b980367c7-kube-api-access-4fbx8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f5b253a6-9fcb-432b-81d9-6a6b980367c7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.183929 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f5b253a6-9fcb-432b-81d9-6a6b980367c7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.184116 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fbx8\" (UniqueName: \"kubernetes.io/projected/f5b253a6-9fcb-432b-81d9-6a6b980367c7-kube-api-access-4fbx8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f5b253a6-9fcb-432b-81d9-6a6b980367c7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.184596 4573 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f5b253a6-9fcb-432b-81d9-6a6b980367c7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.216969 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fbx8\" (UniqueName: \"kubernetes.io/projected/f5b253a6-9fcb-432b-81d9-6a6b980367c7-kube-api-access-4fbx8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f5b253a6-9fcb-432b-81d9-6a6b980367c7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.219506 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f5b253a6-9fcb-432b-81d9-6a6b980367c7\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.273654 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 09:56:10 crc kubenswrapper[4573]: I1203 09:56:10.771379 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 09:56:11 crc kubenswrapper[4573]: I1203 09:56:11.030574 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:56:11 crc kubenswrapper[4573]: E1203 09:56:11.030912 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:56:11 crc kubenswrapper[4573]: I1203 09:56:11.541554 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f5b253a6-9fcb-432b-81d9-6a6b980367c7","Type":"ContainerStarted","Data":"35ac3a64f21139c730a1fc7604c31b2c1b8241be81f32be06600d799e0d57be6"} Dec 03 09:56:12 crc kubenswrapper[4573]: I1203 09:56:12.551015 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f5b253a6-9fcb-432b-81d9-6a6b980367c7","Type":"ContainerStarted","Data":"e26ed3c544d5c6753256d1b220e56f88acbc9d124d5ebbf1ee4cd7ebea26f558"} Dec 03 09:56:25 crc kubenswrapper[4573]: I1203 09:56:25.030576 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:56:25 crc kubenswrapper[4573]: E1203 09:56:25.031452 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.423726 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=25.32453649 podStartE2EDuration="26.423681704s" podCreationTimestamp="2025-12-03 09:56:09 +0000 UTC" firstStartedPulling="2025-12-03 09:56:10.792311376 +0000 UTC m=+4691.360690635" lastFinishedPulling="2025-12-03 09:56:11.891456589 +0000 UTC m=+4692.459835849" observedRunningTime="2025-12-03 09:56:12.567988455 +0000 UTC m=+4693.136367734" watchObservedRunningTime="2025-12-03 09:56:35.423681704 +0000 UTC m=+4715.992060963" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.428744 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vzks7/must-gather-658jg"] Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.431201 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.438658 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vzks7"/"openshift-service-ca.crt" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.438952 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vzks7"/"kube-root-ca.crt" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.449006 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vzks7/must-gather-658jg"] Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.540555 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbxmv\" (UniqueName: \"kubernetes.io/projected/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-kube-api-access-nbxmv\") pod \"must-gather-658jg\" (UID: \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\") " pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.540740 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-must-gather-output\") pod \"must-gather-658jg\" (UID: \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\") " pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.642962 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-must-gather-output\") pod \"must-gather-658jg\" (UID: \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\") " pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.643135 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbxmv\" (UniqueName: \"kubernetes.io/projected/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-kube-api-access-nbxmv\") pod \"must-gather-658jg\" (UID: \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\") " pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.643559 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-must-gather-output\") pod \"must-gather-658jg\" (UID: \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\") " pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.672257 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbxmv\" (UniqueName: \"kubernetes.io/projected/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-kube-api-access-nbxmv\") pod \"must-gather-658jg\" (UID: \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\") " pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 09:56:35 crc kubenswrapper[4573]: I1203 09:56:35.760485 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 09:56:36 crc kubenswrapper[4573]: I1203 09:56:36.434409 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vzks7/must-gather-658jg"] Dec 03 09:56:36 crc kubenswrapper[4573]: I1203 09:56:36.863603 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/must-gather-658jg" event={"ID":"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f","Type":"ContainerStarted","Data":"256a55e5983a79df65d20504cc1b5e3972501bb3d30fc1925421d4af201e9cd5"} Dec 03 09:56:38 crc kubenswrapper[4573]: I1203 09:56:38.032939 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:56:38 crc kubenswrapper[4573]: E1203 09:56:38.033312 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:56:44 crc kubenswrapper[4573]: I1203 09:56:44.953138 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/must-gather-658jg" event={"ID":"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f","Type":"ContainerStarted","Data":"ec9663268efc71d193fe7d5b5b65c8b5f602b2577f73153a7cc4c0469e87cf3b"} Dec 03 09:56:44 crc kubenswrapper[4573]: I1203 09:56:44.953548 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/must-gather-658jg" event={"ID":"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f","Type":"ContainerStarted","Data":"1cf28177f9adbcf0a9c79a12471b1a52b0cd46040c717fee0d7d425a0e157447"} Dec 03 09:56:44 crc kubenswrapper[4573]: I1203 09:56:44.977848 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vzks7/must-gather-658jg" podStartSLOduration=2.549395821 podStartE2EDuration="9.977824164s" podCreationTimestamp="2025-12-03 09:56:35 +0000 UTC" firstStartedPulling="2025-12-03 09:56:36.733311022 +0000 UTC m=+4717.301690281" lastFinishedPulling="2025-12-03 09:56:44.161739365 +0000 UTC m=+4724.730118624" observedRunningTime="2025-12-03 09:56:44.968328046 +0000 UTC m=+4725.536707325" watchObservedRunningTime="2025-12-03 09:56:44.977824164 +0000 UTC m=+4725.546203423" Dec 03 09:56:49 crc kubenswrapper[4573]: I1203 09:56:49.784835 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vzks7/crc-debug-j558f"] Dec 03 09:56:49 crc kubenswrapper[4573]: I1203 09:56:49.787386 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:56:49 crc kubenswrapper[4573]: I1203 09:56:49.789597 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vzks7"/"default-dockercfg-7wbwx" Dec 03 09:56:49 crc kubenswrapper[4573]: I1203 09:56:49.979138 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rwhs\" (UniqueName: \"kubernetes.io/projected/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-kube-api-access-9rwhs\") pod \"crc-debug-j558f\" (UID: \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\") " pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:56:49 crc kubenswrapper[4573]: I1203 09:56:49.979390 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-host\") pod \"crc-debug-j558f\" (UID: \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\") " pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:56:50 crc kubenswrapper[4573]: I1203 09:56:50.081415 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rwhs\" (UniqueName: \"kubernetes.io/projected/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-kube-api-access-9rwhs\") pod \"crc-debug-j558f\" (UID: \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\") " pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:56:50 crc kubenswrapper[4573]: I1203 09:56:50.081567 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-host\") pod \"crc-debug-j558f\" (UID: \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\") " pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:56:50 crc kubenswrapper[4573]: I1203 09:56:50.081686 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-host\") pod \"crc-debug-j558f\" (UID: \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\") " pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:56:50 crc kubenswrapper[4573]: I1203 09:56:50.105949 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rwhs\" (UniqueName: \"kubernetes.io/projected/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-kube-api-access-9rwhs\") pod \"crc-debug-j558f\" (UID: \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\") " pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:56:50 crc kubenswrapper[4573]: I1203 09:56:50.113093 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:56:51 crc kubenswrapper[4573]: I1203 09:56:51.010887 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/crc-debug-j558f" event={"ID":"a7c4a150-c8ac-4e9d-b3cb-0397951952a0","Type":"ContainerStarted","Data":"9fb71e80bf67391b671d1225b00765d0fb4a945baac77bd28b9f99b1f5ddc91f"} Dec 03 09:56:53 crc kubenswrapper[4573]: I1203 09:56:53.031438 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:56:53 crc kubenswrapper[4573]: E1203 09:56:53.031983 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:56:58 crc kubenswrapper[4573]: I1203 09:56:58.724238 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="1999a198-5df7-40fc-ba0f-5b650b3027a6" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 03 09:57:03 crc kubenswrapper[4573]: I1203 09:57:03.128557 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/crc-debug-j558f" event={"ID":"a7c4a150-c8ac-4e9d-b3cb-0397951952a0","Type":"ContainerStarted","Data":"b7931a2ae1256a9b9465a3f3fd6bedf64e7d2e573778be469fb2c9b8d013a1d7"} Dec 03 09:57:03 crc kubenswrapper[4573]: I1203 09:57:03.161020 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vzks7/crc-debug-j558f" podStartSLOduration=2.290188562 podStartE2EDuration="14.161001863s" podCreationTimestamp="2025-12-03 09:56:49 +0000 UTC" firstStartedPulling="2025-12-03 09:56:50.153700232 +0000 UTC m=+4730.722079491" lastFinishedPulling="2025-12-03 09:57:02.024513543 +0000 UTC m=+4742.592892792" observedRunningTime="2025-12-03 09:57:03.157491208 +0000 UTC m=+4743.725870467" watchObservedRunningTime="2025-12-03 09:57:03.161001863 +0000 UTC m=+4743.729381122" Dec 03 09:57:08 crc kubenswrapper[4573]: I1203 09:57:08.037238 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:57:08 crc kubenswrapper[4573]: E1203 09:57:08.039602 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:57:22 crc kubenswrapper[4573]: I1203 09:57:22.031194 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:57:22 crc kubenswrapper[4573]: E1203 09:57:22.033709 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.043799 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bh6t7"] Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.053336 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.062466 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bh6t7"] Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.189523 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-catalog-content\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.189841 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdrtk\" (UniqueName: \"kubernetes.io/projected/67c4d288-b1d5-4e3a-a3f4-0137f9429476-kube-api-access-gdrtk\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.189985 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-utilities\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.291473 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-catalog-content\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.291592 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdrtk\" (UniqueName: \"kubernetes.io/projected/67c4d288-b1d5-4e3a-a3f4-0137f9429476-kube-api-access-gdrtk\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.291673 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-utilities\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.292228 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-catalog-content\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.292327 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-utilities\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.426353 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdrtk\" (UniqueName: \"kubernetes.io/projected/67c4d288-b1d5-4e3a-a3f4-0137f9429476-kube-api-access-gdrtk\") pod \"certified-operators-bh6t7\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:32 crc kubenswrapper[4573]: I1203 09:57:32.687338 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:33 crc kubenswrapper[4573]: I1203 09:57:33.030684 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:57:33 crc kubenswrapper[4573]: E1203 09:57:33.031232 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:57:33 crc kubenswrapper[4573]: I1203 09:57:33.274206 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bh6t7"] Dec 03 09:57:33 crc kubenswrapper[4573]: I1203 09:57:33.438225 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh6t7" event={"ID":"67c4d288-b1d5-4e3a-a3f4-0137f9429476","Type":"ContainerStarted","Data":"e0485d46c378b330b01796efa5787f4fad3a3cdf038999edcb84e759d9a04480"} Dec 03 09:57:34 crc kubenswrapper[4573]: I1203 09:57:34.452326 4573 generic.go:334] "Generic (PLEG): container finished" podID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerID="f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638" exitCode=0 Dec 03 09:57:34 crc kubenswrapper[4573]: I1203 09:57:34.452578 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh6t7" event={"ID":"67c4d288-b1d5-4e3a-a3f4-0137f9429476","Type":"ContainerDied","Data":"f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638"} Dec 03 09:57:36 crc kubenswrapper[4573]: I1203 09:57:36.483242 4573 generic.go:334] "Generic (PLEG): container finished" podID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerID="2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5" exitCode=0 Dec 03 09:57:36 crc kubenswrapper[4573]: I1203 09:57:36.483340 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh6t7" event={"ID":"67c4d288-b1d5-4e3a-a3f4-0137f9429476","Type":"ContainerDied","Data":"2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5"} Dec 03 09:57:37 crc kubenswrapper[4573]: I1203 09:57:37.517539 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh6t7" event={"ID":"67c4d288-b1d5-4e3a-a3f4-0137f9429476","Type":"ContainerStarted","Data":"6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c"} Dec 03 09:57:37 crc kubenswrapper[4573]: I1203 09:57:37.605175 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bh6t7" podStartSLOduration=3.116269899 podStartE2EDuration="5.605149433s" podCreationTimestamp="2025-12-03 09:57:32 +0000 UTC" firstStartedPulling="2025-12-03 09:57:34.454683119 +0000 UTC m=+4775.023062378" lastFinishedPulling="2025-12-03 09:57:36.943562653 +0000 UTC m=+4777.511941912" observedRunningTime="2025-12-03 09:57:37.563998101 +0000 UTC m=+4778.132377360" watchObservedRunningTime="2025-12-03 09:57:37.605149433 +0000 UTC m=+4778.173528682" Dec 03 09:57:42 crc kubenswrapper[4573]: I1203 09:57:42.689101 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:42 crc kubenswrapper[4573]: I1203 09:57:42.689599 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:43 crc kubenswrapper[4573]: I1203 09:57:43.279848 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:43 crc kubenswrapper[4573]: I1203 09:57:43.626895 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:43 crc kubenswrapper[4573]: I1203 09:57:43.723530 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bh6t7"] Dec 03 09:57:45 crc kubenswrapper[4573]: I1203 09:57:45.589537 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bh6t7" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerName="registry-server" containerID="cri-o://6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c" gracePeriod=2 Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.588389 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.599250 4573 generic.go:334] "Generic (PLEG): container finished" podID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerID="6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c" exitCode=0 Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.599296 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh6t7" event={"ID":"67c4d288-b1d5-4e3a-a3f4-0137f9429476","Type":"ContainerDied","Data":"6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c"} Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.599313 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bh6t7" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.599343 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bh6t7" event={"ID":"67c4d288-b1d5-4e3a-a3f4-0137f9429476","Type":"ContainerDied","Data":"e0485d46c378b330b01796efa5787f4fad3a3cdf038999edcb84e759d9a04480"} Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.599374 4573 scope.go:117] "RemoveContainer" containerID="6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.622278 4573 scope.go:117] "RemoveContainer" containerID="2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.664422 4573 scope.go:117] "RemoveContainer" containerID="f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.723769 4573 scope.go:117] "RemoveContainer" containerID="6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c" Dec 03 09:57:46 crc kubenswrapper[4573]: E1203 09:57:46.724263 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c\": container with ID starting with 6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c not found: ID does not exist" containerID="6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.724324 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c"} err="failed to get container status \"6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c\": rpc error: code = NotFound desc = could not find container \"6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c\": container with ID starting with 6dff207bba38ac845d48559b41fc8e233c3d54706b545fcfebddd0d6d8eb989c not found: ID does not exist" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.724375 4573 scope.go:117] "RemoveContainer" containerID="2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5" Dec 03 09:57:46 crc kubenswrapper[4573]: E1203 09:57:46.724738 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5\": container with ID starting with 2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5 not found: ID does not exist" containerID="2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.724770 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5"} err="failed to get container status \"2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5\": rpc error: code = NotFound desc = could not find container \"2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5\": container with ID starting with 2a4b9771dcb47707538ee53b9e0bf3979642226f1b85b8b1e16adf6d7588d8f5 not found: ID does not exist" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.724793 4573 scope.go:117] "RemoveContainer" containerID="f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638" Dec 03 09:57:46 crc kubenswrapper[4573]: E1203 09:57:46.725087 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638\": container with ID starting with f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638 not found: ID does not exist" containerID="f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.725117 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638"} err="failed to get container status \"f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638\": rpc error: code = NotFound desc = could not find container \"f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638\": container with ID starting with f74a0ee3f2c17c5ea6c313ee6cf1b54971dc68e331996180e0d808b4b18e4638 not found: ID does not exist" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.727147 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdrtk\" (UniqueName: \"kubernetes.io/projected/67c4d288-b1d5-4e3a-a3f4-0137f9429476-kube-api-access-gdrtk\") pod \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.727222 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-catalog-content\") pod \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.727394 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-utilities\") pod \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\" (UID: \"67c4d288-b1d5-4e3a-a3f4-0137f9429476\") " Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.728231 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-utilities" (OuterVolumeSpecName: "utilities") pod "67c4d288-b1d5-4e3a-a3f4-0137f9429476" (UID: "67c4d288-b1d5-4e3a-a3f4-0137f9429476"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.770733 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c4d288-b1d5-4e3a-a3f4-0137f9429476-kube-api-access-gdrtk" (OuterVolumeSpecName: "kube-api-access-gdrtk") pod "67c4d288-b1d5-4e3a-a3f4-0137f9429476" (UID: "67c4d288-b1d5-4e3a-a3f4-0137f9429476"). InnerVolumeSpecName "kube-api-access-gdrtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.812821 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67c4d288-b1d5-4e3a-a3f4-0137f9429476" (UID: "67c4d288-b1d5-4e3a-a3f4-0137f9429476"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.831222 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.831261 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdrtk\" (UniqueName: \"kubernetes.io/projected/67c4d288-b1d5-4e3a-a3f4-0137f9429476-kube-api-access-gdrtk\") on node \"crc\" DevicePath \"\"" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.831272 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c4d288-b1d5-4e3a-a3f4-0137f9429476-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.940595 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bh6t7"] Dec 03 09:57:46 crc kubenswrapper[4573]: I1203 09:57:46.962007 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bh6t7"] Dec 03 09:57:48 crc kubenswrapper[4573]: I1203 09:57:48.030121 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:57:48 crc kubenswrapper[4573]: E1203 09:57:48.030845 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:57:48 crc kubenswrapper[4573]: I1203 09:57:48.041037 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" path="/var/lib/kubelet/pods/67c4d288-b1d5-4e3a-a3f4-0137f9429476/volumes" Dec 03 09:57:57 crc kubenswrapper[4573]: I1203 09:57:57.703241 4573 generic.go:334] "Generic (PLEG): container finished" podID="a7c4a150-c8ac-4e9d-b3cb-0397951952a0" containerID="b7931a2ae1256a9b9465a3f3fd6bedf64e7d2e573778be469fb2c9b8d013a1d7" exitCode=0 Dec 03 09:57:57 crc kubenswrapper[4573]: I1203 09:57:57.703345 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/crc-debug-j558f" event={"ID":"a7c4a150-c8ac-4e9d-b3cb-0397951952a0","Type":"ContainerDied","Data":"b7931a2ae1256a9b9465a3f3fd6bedf64e7d2e573778be469fb2c9b8d013a1d7"} Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.827134 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.864978 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vzks7/crc-debug-j558f"] Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.872383 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vzks7/crc-debug-j558f"] Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.882918 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-host\") pod \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\" (UID: \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\") " Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.883171 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rwhs\" (UniqueName: \"kubernetes.io/projected/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-kube-api-access-9rwhs\") pod \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\" (UID: \"a7c4a150-c8ac-4e9d-b3cb-0397951952a0\") " Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.883254 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-host" (OuterVolumeSpecName: "host") pod "a7c4a150-c8ac-4e9d-b3cb-0397951952a0" (UID: "a7c4a150-c8ac-4e9d-b3cb-0397951952a0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.884030 4573 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-host\") on node \"crc\" DevicePath \"\"" Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.891268 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-kube-api-access-9rwhs" (OuterVolumeSpecName: "kube-api-access-9rwhs") pod "a7c4a150-c8ac-4e9d-b3cb-0397951952a0" (UID: "a7c4a150-c8ac-4e9d-b3cb-0397951952a0"). InnerVolumeSpecName "kube-api-access-9rwhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:57:58 crc kubenswrapper[4573]: I1203 09:57:58.985325 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rwhs\" (UniqueName: \"kubernetes.io/projected/a7c4a150-c8ac-4e9d-b3cb-0397951952a0-kube-api-access-9rwhs\") on node \"crc\" DevicePath \"\"" Dec 03 09:57:59 crc kubenswrapper[4573]: I1203 09:57:59.722155 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fb71e80bf67391b671d1225b00765d0fb4a945baac77bd28b9f99b1f5ddc91f" Dec 03 09:57:59 crc kubenswrapper[4573]: I1203 09:57:59.722512 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-j558f" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.070831 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7c4a150-c8ac-4e9d-b3cb-0397951952a0" path="/var/lib/kubelet/pods/a7c4a150-c8ac-4e9d-b3cb-0397951952a0/volumes" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.121498 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vzks7/crc-debug-fg6s8"] Dec 03 09:58:00 crc kubenswrapper[4573]: E1203 09:58:00.121909 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerName="extract-content" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.121927 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerName="extract-content" Dec 03 09:58:00 crc kubenswrapper[4573]: E1203 09:58:00.121942 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerName="registry-server" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.121948 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerName="registry-server" Dec 03 09:58:00 crc kubenswrapper[4573]: E1203 09:58:00.121965 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerName="extract-utilities" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.121971 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerName="extract-utilities" Dec 03 09:58:00 crc kubenswrapper[4573]: E1203 09:58:00.121986 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7c4a150-c8ac-4e9d-b3cb-0397951952a0" containerName="container-00" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.121991 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7c4a150-c8ac-4e9d-b3cb-0397951952a0" containerName="container-00" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.122203 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c4d288-b1d5-4e3a-a3f4-0137f9429476" containerName="registry-server" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.122227 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7c4a150-c8ac-4e9d-b3cb-0397951952a0" containerName="container-00" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.122823 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.125776 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vzks7"/"default-dockercfg-7wbwx" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.134143 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj97n\" (UniqueName: \"kubernetes.io/projected/8a1151d8-458b-4925-a687-250615d012a2-kube-api-access-kj97n\") pod \"crc-debug-fg6s8\" (UID: \"8a1151d8-458b-4925-a687-250615d012a2\") " pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.134305 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a1151d8-458b-4925-a687-250615d012a2-host\") pod \"crc-debug-fg6s8\" (UID: \"8a1151d8-458b-4925-a687-250615d012a2\") " pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.235838 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj97n\" (UniqueName: \"kubernetes.io/projected/8a1151d8-458b-4925-a687-250615d012a2-kube-api-access-kj97n\") pod \"crc-debug-fg6s8\" (UID: \"8a1151d8-458b-4925-a687-250615d012a2\") " pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.235935 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a1151d8-458b-4925-a687-250615d012a2-host\") pod \"crc-debug-fg6s8\" (UID: \"8a1151d8-458b-4925-a687-250615d012a2\") " pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.236162 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a1151d8-458b-4925-a687-250615d012a2-host\") pod \"crc-debug-fg6s8\" (UID: \"8a1151d8-458b-4925-a687-250615d012a2\") " pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.255783 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj97n\" (UniqueName: \"kubernetes.io/projected/8a1151d8-458b-4925-a687-250615d012a2-kube-api-access-kj97n\") pod \"crc-debug-fg6s8\" (UID: \"8a1151d8-458b-4925-a687-250615d012a2\") " pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.446041 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.734852 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/crc-debug-fg6s8" event={"ID":"8a1151d8-458b-4925-a687-250615d012a2","Type":"ContainerStarted","Data":"8fe94cd29db6d826f6adb25629c474d5e69d3a400b2b5aabbd61e6215f0d5221"} Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.735166 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/crc-debug-fg6s8" event={"ID":"8a1151d8-458b-4925-a687-250615d012a2","Type":"ContainerStarted","Data":"ccafc2f0d76ed063b82bd6824d1c931b3f6afb771f053615fba5f96e6555204f"} Dec 03 09:58:00 crc kubenswrapper[4573]: I1203 09:58:00.756794 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vzks7/crc-debug-fg6s8" podStartSLOduration=0.756753778 podStartE2EDuration="756.753778ms" podCreationTimestamp="2025-12-03 09:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 09:58:00.747143306 +0000 UTC m=+4801.315522565" watchObservedRunningTime="2025-12-03 09:58:00.756753778 +0000 UTC m=+4801.325133047" Dec 03 09:58:01 crc kubenswrapper[4573]: I1203 09:58:01.762855 4573 generic.go:334] "Generic (PLEG): container finished" podID="8a1151d8-458b-4925-a687-250615d012a2" containerID="8fe94cd29db6d826f6adb25629c474d5e69d3a400b2b5aabbd61e6215f0d5221" exitCode=0 Dec 03 09:58:01 crc kubenswrapper[4573]: I1203 09:58:01.762936 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/crc-debug-fg6s8" event={"ID":"8a1151d8-458b-4925-a687-250615d012a2","Type":"ContainerDied","Data":"8fe94cd29db6d826f6adb25629c474d5e69d3a400b2b5aabbd61e6215f0d5221"} Dec 03 09:58:02 crc kubenswrapper[4573]: I1203 09:58:02.030886 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:58:02 crc kubenswrapper[4573]: E1203 09:58:02.031145 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.040789 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.113355 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a1151d8-458b-4925-a687-250615d012a2-host\") pod \"8a1151d8-458b-4925-a687-250615d012a2\" (UID: \"8a1151d8-458b-4925-a687-250615d012a2\") " Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.113502 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a1151d8-458b-4925-a687-250615d012a2-host" (OuterVolumeSpecName: "host") pod "8a1151d8-458b-4925-a687-250615d012a2" (UID: "8a1151d8-458b-4925-a687-250615d012a2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.113757 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj97n\" (UniqueName: \"kubernetes.io/projected/8a1151d8-458b-4925-a687-250615d012a2-kube-api-access-kj97n\") pod \"8a1151d8-458b-4925-a687-250615d012a2\" (UID: \"8a1151d8-458b-4925-a687-250615d012a2\") " Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.114273 4573 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a1151d8-458b-4925-a687-250615d012a2-host\") on node \"crc\" DevicePath \"\"" Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.122212 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a1151d8-458b-4925-a687-250615d012a2-kube-api-access-kj97n" (OuterVolumeSpecName: "kube-api-access-kj97n") pod "8a1151d8-458b-4925-a687-250615d012a2" (UID: "8a1151d8-458b-4925-a687-250615d012a2"). InnerVolumeSpecName "kube-api-access-kj97n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.215583 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj97n\" (UniqueName: \"kubernetes.io/projected/8a1151d8-458b-4925-a687-250615d012a2-kube-api-access-kj97n\") on node \"crc\" DevicePath \"\"" Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.376794 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vzks7/crc-debug-fg6s8"] Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.386801 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vzks7/crc-debug-fg6s8"] Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.784531 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccafc2f0d76ed063b82bd6824d1c931b3f6afb771f053615fba5f96e6555204f" Dec 03 09:58:03 crc kubenswrapper[4573]: I1203 09:58:03.784630 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-fg6s8" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.048197 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a1151d8-458b-4925-a687-250615d012a2" path="/var/lib/kubelet/pods/8a1151d8-458b-4925-a687-250615d012a2/volumes" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.638220 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vzks7/crc-debug-5d64s"] Dec 03 09:58:04 crc kubenswrapper[4573]: E1203 09:58:04.638792 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a1151d8-458b-4925-a687-250615d012a2" containerName="container-00" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.638814 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a1151d8-458b-4925-a687-250615d012a2" containerName="container-00" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.639133 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a1151d8-458b-4925-a687-250615d012a2" containerName="container-00" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.640008 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.644727 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vzks7"/"default-dockercfg-7wbwx" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.745841 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-host\") pod \"crc-debug-5d64s\" (UID: \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\") " pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.745950 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp7zm\" (UniqueName: \"kubernetes.io/projected/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-kube-api-access-jp7zm\") pod \"crc-debug-5d64s\" (UID: \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\") " pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.846932 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp7zm\" (UniqueName: \"kubernetes.io/projected/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-kube-api-access-jp7zm\") pod \"crc-debug-5d64s\" (UID: \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\") " pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.847076 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-host\") pod \"crc-debug-5d64s\" (UID: \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\") " pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:04 crc kubenswrapper[4573]: I1203 09:58:04.847154 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-host\") pod \"crc-debug-5d64s\" (UID: \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\") " pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:05 crc kubenswrapper[4573]: I1203 09:58:05.826241 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp7zm\" (UniqueName: \"kubernetes.io/projected/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-kube-api-access-jp7zm\") pod \"crc-debug-5d64s\" (UID: \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\") " pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:05 crc kubenswrapper[4573]: I1203 09:58:05.863116 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:06 crc kubenswrapper[4573]: I1203 09:58:06.812476 4573 generic.go:334] "Generic (PLEG): container finished" podID="e0900e1f-d595-42a0-88b1-cbfccdf82dfa" containerID="11175747986d5928252b4acd1a10f657f8623f8fa8cf0cf5c463495e38156bc7" exitCode=0 Dec 03 09:58:06 crc kubenswrapper[4573]: I1203 09:58:06.812551 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/crc-debug-5d64s" event={"ID":"e0900e1f-d595-42a0-88b1-cbfccdf82dfa","Type":"ContainerDied","Data":"11175747986d5928252b4acd1a10f657f8623f8fa8cf0cf5c463495e38156bc7"} Dec 03 09:58:06 crc kubenswrapper[4573]: I1203 09:58:06.813018 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/crc-debug-5d64s" event={"ID":"e0900e1f-d595-42a0-88b1-cbfccdf82dfa","Type":"ContainerStarted","Data":"fc4d6b97bd33610e6a4c12d3fbdb2a9f36c632ec45c8f31a909e7bf99dca0c68"} Dec 03 09:58:06 crc kubenswrapper[4573]: I1203 09:58:06.869016 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vzks7/crc-debug-5d64s"] Dec 03 09:58:06 crc kubenswrapper[4573]: I1203 09:58:06.882753 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vzks7/crc-debug-5d64s"] Dec 03 09:58:07 crc kubenswrapper[4573]: I1203 09:58:07.958298 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:08 crc kubenswrapper[4573]: I1203 09:58:08.118612 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp7zm\" (UniqueName: \"kubernetes.io/projected/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-kube-api-access-jp7zm\") pod \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\" (UID: \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\") " Dec 03 09:58:08 crc kubenswrapper[4573]: I1203 09:58:08.118913 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-host\") pod \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\" (UID: \"e0900e1f-d595-42a0-88b1-cbfccdf82dfa\") " Dec 03 09:58:08 crc kubenswrapper[4573]: I1203 09:58:08.119396 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-host" (OuterVolumeSpecName: "host") pod "e0900e1f-d595-42a0-88b1-cbfccdf82dfa" (UID: "e0900e1f-d595-42a0-88b1-cbfccdf82dfa"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 09:58:08 crc kubenswrapper[4573]: I1203 09:58:08.126443 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-kube-api-access-jp7zm" (OuterVolumeSpecName: "kube-api-access-jp7zm") pod "e0900e1f-d595-42a0-88b1-cbfccdf82dfa" (UID: "e0900e1f-d595-42a0-88b1-cbfccdf82dfa"). InnerVolumeSpecName "kube-api-access-jp7zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:58:08 crc kubenswrapper[4573]: I1203 09:58:08.222512 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp7zm\" (UniqueName: \"kubernetes.io/projected/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-kube-api-access-jp7zm\") on node \"crc\" DevicePath \"\"" Dec 03 09:58:08 crc kubenswrapper[4573]: I1203 09:58:08.222554 4573 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e0900e1f-d595-42a0-88b1-cbfccdf82dfa-host\") on node \"crc\" DevicePath \"\"" Dec 03 09:58:08 crc kubenswrapper[4573]: I1203 09:58:08.836785 4573 scope.go:117] "RemoveContainer" containerID="11175747986d5928252b4acd1a10f657f8623f8fa8cf0cf5c463495e38156bc7" Dec 03 09:58:08 crc kubenswrapper[4573]: I1203 09:58:08.836934 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/crc-debug-5d64s" Dec 03 09:58:10 crc kubenswrapper[4573]: I1203 09:58:10.042570 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0900e1f-d595-42a0-88b1-cbfccdf82dfa" path="/var/lib/kubelet/pods/e0900e1f-d595-42a0-88b1-cbfccdf82dfa/volumes" Dec 03 09:58:15 crc kubenswrapper[4573]: I1203 09:58:15.030707 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:58:15 crc kubenswrapper[4573]: E1203 09:58:15.031563 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.039523 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:58:30 crc kubenswrapper[4573]: E1203 09:58:30.040305 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.186171 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-99whl"] Dec 03 09:58:30 crc kubenswrapper[4573]: E1203 09:58:30.186582 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0900e1f-d595-42a0-88b1-cbfccdf82dfa" containerName="container-00" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.186600 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0900e1f-d595-42a0-88b1-cbfccdf82dfa" containerName="container-00" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.186795 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0900e1f-d595-42a0-88b1-cbfccdf82dfa" containerName="container-00" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.188100 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.207104 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-99whl"] Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.368073 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppqqp\" (UniqueName: \"kubernetes.io/projected/aed85f45-6939-4da3-b93d-29d48d3dc433-kube-api-access-ppqqp\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.368444 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-utilities\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.368639 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-catalog-content\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.470648 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppqqp\" (UniqueName: \"kubernetes.io/projected/aed85f45-6939-4da3-b93d-29d48d3dc433-kube-api-access-ppqqp\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.470793 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-utilities\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.470873 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-catalog-content\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.471377 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-catalog-content\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.471916 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-utilities\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.492205 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppqqp\" (UniqueName: \"kubernetes.io/projected/aed85f45-6939-4da3-b93d-29d48d3dc433-kube-api-access-ppqqp\") pod \"redhat-marketplace-99whl\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.517121 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:30 crc kubenswrapper[4573]: I1203 09:58:30.995799 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-99whl"] Dec 03 09:58:31 crc kubenswrapper[4573]: I1203 09:58:31.042228 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-99whl" event={"ID":"aed85f45-6939-4da3-b93d-29d48d3dc433","Type":"ContainerStarted","Data":"938e3fc3d8b5de172c4c213f1b4b502245277285cbb6afb2f7a9bf64ec18eb84"} Dec 03 09:58:32 crc kubenswrapper[4573]: I1203 09:58:32.066127 4573 generic.go:334] "Generic (PLEG): container finished" podID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerID="c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52" exitCode=0 Dec 03 09:58:32 crc kubenswrapper[4573]: I1203 09:58:32.066910 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-99whl" event={"ID":"aed85f45-6939-4da3-b93d-29d48d3dc433","Type":"ContainerDied","Data":"c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52"} Dec 03 09:58:32 crc kubenswrapper[4573]: I1203 09:58:32.069184 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 09:58:34 crc kubenswrapper[4573]: I1203 09:58:34.089590 4573 generic.go:334] "Generic (PLEG): container finished" podID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerID="42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a" exitCode=0 Dec 03 09:58:34 crc kubenswrapper[4573]: I1203 09:58:34.089629 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-99whl" event={"ID":"aed85f45-6939-4da3-b93d-29d48d3dc433","Type":"ContainerDied","Data":"42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a"} Dec 03 09:58:35 crc kubenswrapper[4573]: I1203 09:58:35.504062 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-847b4dbb6-dsqmk_6907cde0-9a5e-432e-bc68-55b357a514d9/barbican-api/0.log" Dec 03 09:58:35 crc kubenswrapper[4573]: I1203 09:58:35.748143 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-847b4dbb6-dsqmk_6907cde0-9a5e-432e-bc68-55b357a514d9/barbican-api-log/0.log" Dec 03 09:58:35 crc kubenswrapper[4573]: I1203 09:58:35.804024 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c965d87c4-dmtfc_e954edc8-66c8-4f8f-a255-b8c5d810aa1a/barbican-keystone-listener/0.log" Dec 03 09:58:35 crc kubenswrapper[4573]: I1203 09:58:35.921613 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c965d87c4-dmtfc_e954edc8-66c8-4f8f-a255-b8c5d810aa1a/barbican-keystone-listener-log/0.log" Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.117928 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-99whl" event={"ID":"aed85f45-6939-4da3-b93d-29d48d3dc433","Type":"ContainerStarted","Data":"d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf"} Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.147683 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-99whl" podStartSLOduration=3.252526299 podStartE2EDuration="6.147664095s" podCreationTimestamp="2025-12-03 09:58:30 +0000 UTC" firstStartedPulling="2025-12-03 09:58:32.068858673 +0000 UTC m=+4832.637237932" lastFinishedPulling="2025-12-03 09:58:34.963996479 +0000 UTC m=+4835.532375728" observedRunningTime="2025-12-03 09:58:36.141531818 +0000 UTC m=+4836.709911077" watchObservedRunningTime="2025-12-03 09:58:36.147664095 +0000 UTC m=+4836.716043354" Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.168162 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-cd9d86d7-kn6nh_865e87dd-1725-4932-9566-dbfbf7b85a60/barbican-worker/0.log" Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.179633 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-cd9d86d7-kn6nh_865e87dd-1725-4932-9566-dbfbf7b85a60/barbican-worker-log/0.log" Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.610532 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r_3dc88703-275c-419a-961e-9c034464b6cb/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.704801 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3601284e-acb1-4e60-a4c1-18c404dc4072/ceilometer-central-agent/0.log" Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.805940 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3601284e-acb1-4e60-a4c1-18c404dc4072/ceilometer-notification-agent/0.log" Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.884175 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3601284e-acb1-4e60-a4c1-18c404dc4072/sg-core/0.log" Dec 03 09:58:36 crc kubenswrapper[4573]: I1203 09:58:36.917898 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3601284e-acb1-4e60-a4c1-18c404dc4072/proxy-httpd/0.log" Dec 03 09:58:37 crc kubenswrapper[4573]: I1203 09:58:37.121142 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3f75b291-5f0e-422f-a640-a3a231c400ee/cinder-api/0.log" Dec 03 09:58:37 crc kubenswrapper[4573]: I1203 09:58:37.159431 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3f75b291-5f0e-422f-a640-a3a231c400ee/cinder-api-log/0.log" Dec 03 09:58:37 crc kubenswrapper[4573]: I1203 09:58:37.525105 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_20097614-b580-49bf-a8c7-04f371825b9a/cinder-scheduler/0.log" Dec 03 09:58:37 crc kubenswrapper[4573]: I1203 09:58:37.528940 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8_ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:37 crc kubenswrapper[4573]: I1203 09:58:37.530401 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_20097614-b580-49bf-a8c7-04f371825b9a/probe/0.log" Dec 03 09:58:37 crc kubenswrapper[4573]: I1203 09:58:37.798930 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-swdkm_6ea550b6-a03c-42ff-b766-073b5ca7bf15/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:37 crc kubenswrapper[4573]: I1203 09:58:37.887334 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d47554775-5hw5x_eb528c46-bf42-4870-850c-dcf96e5a59b9/init/0.log" Dec 03 09:58:38 crc kubenswrapper[4573]: I1203 09:58:38.049212 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d47554775-5hw5x_eb528c46-bf42-4870-850c-dcf96e5a59b9/init/0.log" Dec 03 09:58:38 crc kubenswrapper[4573]: I1203 09:58:38.150402 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-6kp67_960face3-051e-4d24-b7ae-8e8b0f13d14e/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:38 crc kubenswrapper[4573]: I1203 09:58:38.267269 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d47554775-5hw5x_eb528c46-bf42-4870-850c-dcf96e5a59b9/dnsmasq-dns/0.log" Dec 03 09:58:38 crc kubenswrapper[4573]: I1203 09:58:38.391718 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_36bf5497-62bd-4e75-99e8-0a0272cd644a/glance-log/0.log" Dec 03 09:58:38 crc kubenswrapper[4573]: I1203 09:58:38.430699 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_36bf5497-62bd-4e75-99e8-0a0272cd644a/glance-httpd/0.log" Dec 03 09:58:38 crc kubenswrapper[4573]: I1203 09:58:38.640451 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2/glance-log/0.log" Dec 03 09:58:38 crc kubenswrapper[4573]: I1203 09:58:38.750981 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2/glance-httpd/0.log" Dec 03 09:58:39 crc kubenswrapper[4573]: I1203 09:58:39.395622 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-66f99c579b-gh7sh_b78e2453-adc3-4139-97c5-9f8a977df1ee/horizon/2.log" Dec 03 09:58:39 crc kubenswrapper[4573]: I1203 09:58:39.409678 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p_66bd3982-db80-45dc-a985-5b2f05a0e257/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:39 crc kubenswrapper[4573]: I1203 09:58:39.538230 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-66f99c579b-gh7sh_b78e2453-adc3-4139-97c5-9f8a977df1ee/horizon/1.log" Dec 03 09:58:39 crc kubenswrapper[4573]: I1203 09:58:39.946700 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-cg5z8_2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:40 crc kubenswrapper[4573]: I1203 09:58:40.263593 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_e50ae89c-c3c5-44a5-9222-b14ddbadc294/kube-state-metrics/0.log" Dec 03 09:58:40 crc kubenswrapper[4573]: I1203 09:58:40.402330 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-66f99c579b-gh7sh_b78e2453-adc3-4139-97c5-9f8a977df1ee/horizon-log/0.log" Dec 03 09:58:40 crc kubenswrapper[4573]: I1203 09:58:40.516177 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d_022b0b02-d8bb-4560-af50-5d048f4ec520/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:40 crc kubenswrapper[4573]: I1203 09:58:40.518155 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:40 crc kubenswrapper[4573]: I1203 09:58:40.520147 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:40 crc kubenswrapper[4573]: I1203 09:58:40.586360 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:40 crc kubenswrapper[4573]: I1203 09:58:40.856949 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-f56bd7d55-5qgxb_06a92024-3b18-43c5-a4aa-22cec1d43ff3/keystone-api/0.log" Dec 03 09:58:41 crc kubenswrapper[4573]: I1203 09:58:41.236497 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:41 crc kubenswrapper[4573]: I1203 09:58:41.289153 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-99whl"] Dec 03 09:58:41 crc kubenswrapper[4573]: I1203 09:58:41.450988 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cfccd4555-2shp6_4434273d-ce2e-4ecd-820d-7f2ecf35bca7/neutron-api/0.log" Dec 03 09:58:41 crc kubenswrapper[4573]: I1203 09:58:41.538855 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44_281980d3-9cd5-4f71-b6a1-8602818bc836/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:41 crc kubenswrapper[4573]: I1203 09:58:41.582147 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cfccd4555-2shp6_4434273d-ce2e-4ecd-820d-7f2ecf35bca7/neutron-httpd/0.log" Dec 03 09:58:42 crc kubenswrapper[4573]: I1203 09:58:42.030124 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:58:42 crc kubenswrapper[4573]: E1203 09:58:42.030365 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:58:42 crc kubenswrapper[4573]: I1203 09:58:42.453568 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a96a7087-9cec-427a-b088-0d900045bde3/nova-cell0-conductor-conductor/0.log" Dec 03 09:58:42 crc kubenswrapper[4573]: I1203 09:58:42.651109 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2227cd84-de1e-4b88-8712-dbd6471e6c67/nova-cell1-conductor-conductor/0.log" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.080846 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_192989bd-f8d0-43d2-9267-3022e451041c/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.195440 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-99whl" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerName="registry-server" containerID="cri-o://d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf" gracePeriod=2 Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.313467 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b06177f1-1c12-41d9-b7f8-7df5e6986872/nova-api-log/0.log" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.460002 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-jc6rt_5847f67b-3130-4fa9-831f-bcfe0c27ad48/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.603367 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b06177f1-1c12-41d9-b7f8-7df5e6986872/nova-api-api/0.log" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.734247 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.847551 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-catalog-content\") pod \"aed85f45-6939-4da3-b93d-29d48d3dc433\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.847856 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-utilities\") pod \"aed85f45-6939-4da3-b93d-29d48d3dc433\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.847981 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppqqp\" (UniqueName: \"kubernetes.io/projected/aed85f45-6939-4da3-b93d-29d48d3dc433-kube-api-access-ppqqp\") pod \"aed85f45-6939-4da3-b93d-29d48d3dc433\" (UID: \"aed85f45-6939-4da3-b93d-29d48d3dc433\") " Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.851687 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-utilities" (OuterVolumeSpecName: "utilities") pod "aed85f45-6939-4da3-b93d-29d48d3dc433" (UID: "aed85f45-6939-4da3-b93d-29d48d3dc433"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.868471 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f233cf9b-b14d-4eee-8f84-3fcb5d85f571/nova-metadata-log/0.log" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.879093 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed85f45-6939-4da3-b93d-29d48d3dc433-kube-api-access-ppqqp" (OuterVolumeSpecName: "kube-api-access-ppqqp") pod "aed85f45-6939-4da3-b93d-29d48d3dc433" (UID: "aed85f45-6939-4da3-b93d-29d48d3dc433"). InnerVolumeSpecName "kube-api-access-ppqqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.907154 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aed85f45-6939-4da3-b93d-29d48d3dc433" (UID: "aed85f45-6939-4da3-b93d-29d48d3dc433"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.950024 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.950071 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed85f45-6939-4da3-b93d-29d48d3dc433-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 09:58:43 crc kubenswrapper[4573]: I1203 09:58:43.950083 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppqqp\" (UniqueName: \"kubernetes.io/projected/aed85f45-6939-4da3-b93d-29d48d3dc433-kube-api-access-ppqqp\") on node \"crc\" DevicePath \"\"" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.207335 4573 generic.go:334] "Generic (PLEG): container finished" podID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerID="d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf" exitCode=0 Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.207384 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-99whl" event={"ID":"aed85f45-6939-4da3-b93d-29d48d3dc433","Type":"ContainerDied","Data":"d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf"} Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.207413 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-99whl" event={"ID":"aed85f45-6939-4da3-b93d-29d48d3dc433","Type":"ContainerDied","Data":"938e3fc3d8b5de172c4c213f1b4b502245277285cbb6afb2f7a9bf64ec18eb84"} Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.207432 4573 scope.go:117] "RemoveContainer" containerID="d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.207481 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-99whl" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.248282 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-99whl"] Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.254293 4573 scope.go:117] "RemoveContainer" containerID="42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.260417 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-99whl"] Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.268568 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d/mysql-bootstrap/0.log" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.311212 4573 scope.go:117] "RemoveContainer" containerID="c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.348840 4573 scope.go:117] "RemoveContainer" containerID="d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf" Dec 03 09:58:44 crc kubenswrapper[4573]: E1203 09:58:44.349761 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf\": container with ID starting with d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf not found: ID does not exist" containerID="d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.351182 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf"} err="failed to get container status \"d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf\": rpc error: code = NotFound desc = could not find container \"d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf\": container with ID starting with d7fb0a9be7e2c38b16336bf1d4d5771aaacf9f07b9691dcc91926ff4e43e30cf not found: ID does not exist" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.351435 4573 scope.go:117] "RemoveContainer" containerID="42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a" Dec 03 09:58:44 crc kubenswrapper[4573]: E1203 09:58:44.352275 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a\": container with ID starting with 42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a not found: ID does not exist" containerID="42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.352308 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a"} err="failed to get container status \"42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a\": rpc error: code = NotFound desc = could not find container \"42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a\": container with ID starting with 42264f3be6957d0e591c3a60ff67192b7749ca470b3b1a28c6d84a983e57670a not found: ID does not exist" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.352329 4573 scope.go:117] "RemoveContainer" containerID="c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52" Dec 03 09:58:44 crc kubenswrapper[4573]: E1203 09:58:44.353134 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52\": container with ID starting with c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52 not found: ID does not exist" containerID="c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.353541 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52"} err="failed to get container status \"c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52\": rpc error: code = NotFound desc = could not find container \"c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52\": container with ID starting with c05dbd1527e1d16da67f8fcf8c0e3174a87096f4fb8ad2d320dd4d51b8705a52 not found: ID does not exist" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.505782 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d/mysql-bootstrap/0.log" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.636661 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d/galera/0.log" Dec 03 09:58:44 crc kubenswrapper[4573]: I1203 09:58:44.758934 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_56f569d8-50e2-4177-bec8-8225de477357/nova-scheduler-scheduler/0.log" Dec 03 09:58:45 crc kubenswrapper[4573]: I1203 09:58:45.158879 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96/mysql-bootstrap/0.log" Dec 03 09:58:45 crc kubenswrapper[4573]: I1203 09:58:45.160808 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96/mysql-bootstrap/0.log" Dec 03 09:58:45 crc kubenswrapper[4573]: I1203 09:58:45.192517 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96/galera/0.log" Dec 03 09:58:45 crc kubenswrapper[4573]: I1203 09:58:45.521184 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_4357c5a1-7325-4bff-8a79-4ffd70879b1c/openstackclient/0.log" Dec 03 09:58:45 crc kubenswrapper[4573]: I1203 09:58:45.835672 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bj74b_33a7f303-d65a-4546-bbb5-1e223d48d847/ovn-controller/0.log" Dec 03 09:58:45 crc kubenswrapper[4573]: I1203 09:58:45.894584 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rrcpn_6a19cac1-e9f7-4f7d-b356-3263c1b8c446/openstack-network-exporter/0.log" Dec 03 09:58:45 crc kubenswrapper[4573]: I1203 09:58:45.968981 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f233cf9b-b14d-4eee-8f84-3fcb5d85f571/nova-metadata-metadata/0.log" Dec 03 09:58:46 crc kubenswrapper[4573]: I1203 09:58:46.055065 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" path="/var/lib/kubelet/pods/aed85f45-6939-4da3-b93d-29d48d3dc433/volumes" Dec 03 09:58:46 crc kubenswrapper[4573]: I1203 09:58:46.127656 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jbx79_651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8/ovsdb-server-init/0.log" Dec 03 09:58:46 crc kubenswrapper[4573]: I1203 09:58:46.571519 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jbx79_651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8/ovs-vswitchd/0.log" Dec 03 09:58:46 crc kubenswrapper[4573]: I1203 09:58:46.589764 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jbx79_651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8/ovsdb-server/0.log" Dec 03 09:58:46 crc kubenswrapper[4573]: I1203 09:58:46.602199 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jbx79_651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8/ovsdb-server-init/0.log" Dec 03 09:58:47 crc kubenswrapper[4573]: I1203 09:58:47.329336 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1999a198-5df7-40fc-ba0f-5b650b3027a6/ovn-northd/0.log" Dec 03 09:58:47 crc kubenswrapper[4573]: I1203 09:58:47.361437 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-vq5qw_633c66a3-1e25-4f81-9eee-2cb09bf54855/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:47 crc kubenswrapper[4573]: I1203 09:58:47.447998 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1999a198-5df7-40fc-ba0f-5b650b3027a6/openstack-network-exporter/0.log" Dec 03 09:58:47 crc kubenswrapper[4573]: I1203 09:58:47.610222 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1604082b-b77f-4df2-bea3-ab7803b3e5ec/openstack-network-exporter/0.log" Dec 03 09:58:47 crc kubenswrapper[4573]: I1203 09:58:47.617314 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1604082b-b77f-4df2-bea3-ab7803b3e5ec/ovsdbserver-nb/0.log" Dec 03 09:58:47 crc kubenswrapper[4573]: I1203 09:58:47.869979 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f5af474c-26c9-4877-989a-911cd77862ab/openstack-network-exporter/0.log" Dec 03 09:58:47 crc kubenswrapper[4573]: I1203 09:58:47.917874 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f5af474c-26c9-4877-989a-911cd77862ab/ovsdbserver-sb/0.log" Dec 03 09:58:48 crc kubenswrapper[4573]: I1203 09:58:48.428328 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-85bbcf664-5fm4n_b292e81d-5ded-4fff-bb9e-c764c395609b/placement-api/0.log" Dec 03 09:58:48 crc kubenswrapper[4573]: I1203 09:58:48.452512 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-85bbcf664-5fm4n_b292e81d-5ded-4fff-bb9e-c764c395609b/placement-log/0.log" Dec 03 09:58:48 crc kubenswrapper[4573]: I1203 09:58:48.584466 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f038afb-2485-4e6e-8e0c-99ec178f4268/setup-container/0.log" Dec 03 09:58:48 crc kubenswrapper[4573]: I1203 09:58:48.961602 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f038afb-2485-4e6e-8e0c-99ec178f4268/rabbitmq/0.log" Dec 03 09:58:48 crc kubenswrapper[4573]: I1203 09:58:48.988655 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f038afb-2485-4e6e-8e0c-99ec178f4268/setup-container/0.log" Dec 03 09:58:49 crc kubenswrapper[4573]: I1203 09:58:49.062449 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fb1808-0ac8-4096-abff-76c44b1245f8/setup-container/0.log" Dec 03 09:58:49 crc kubenswrapper[4573]: I1203 09:58:49.263243 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fb1808-0ac8-4096-abff-76c44b1245f8/setup-container/0.log" Dec 03 09:58:49 crc kubenswrapper[4573]: I1203 09:58:49.352637 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fb1808-0ac8-4096-abff-76c44b1245f8/rabbitmq/0.log" Dec 03 09:58:49 crc kubenswrapper[4573]: I1203 09:58:49.388276 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp_fb9f04b9-8dd3-4776-8dec-c469a94759c2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:49 crc kubenswrapper[4573]: I1203 09:58:49.647411 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-lbctc_609d7944-6cf6-47c4-a1f2-2437c57703d2/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:49 crc kubenswrapper[4573]: I1203 09:58:49.700761 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7_dd487e4c-80de-48b3-aaeb-1a3c9d5eda74/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:50 crc kubenswrapper[4573]: I1203 09:58:50.007874 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-qnv8j_fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:50 crc kubenswrapper[4573]: I1203 09:58:50.197505 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-gc8nt_9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3/ssh-known-hosts-edpm-deployment/0.log" Dec 03 09:58:50 crc kubenswrapper[4573]: I1203 09:58:50.501998 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-66d678ddfc-gvbk5_02bc648a-d9c8-4fdf-bf48-bc57a6b05386/proxy-server/0.log" Dec 03 09:58:50 crc kubenswrapper[4573]: I1203 09:58:50.572275 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-66d678ddfc-gvbk5_02bc648a-d9c8-4fdf-bf48-bc57a6b05386/proxy-httpd/0.log" Dec 03 09:58:50 crc kubenswrapper[4573]: I1203 09:58:50.613506 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-lzbxl_5ceed462-6691-4229-b764-b25eb42bbf86/swift-ring-rebalance/0.log" Dec 03 09:58:50 crc kubenswrapper[4573]: I1203 09:58:50.935772 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/account-auditor/0.log" Dec 03 09:58:50 crc kubenswrapper[4573]: I1203 09:58:50.939811 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/account-replicator/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.013778 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/account-reaper/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.256219 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/container-auditor/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.281626 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/account-server/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.322464 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/container-server/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.327494 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/container-replicator/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.524687 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/container-updater/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.641138 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-replicator/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.699173 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-auditor/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.701720 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-expirer/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.928855 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-updater/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.965401 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/rsync/0.log" Dec 03 09:58:51 crc kubenswrapper[4573]: I1203 09:58:51.975262 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-server/0.log" Dec 03 09:58:52 crc kubenswrapper[4573]: I1203 09:58:52.041367 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/swift-recon-cron/0.log" Dec 03 09:58:52 crc kubenswrapper[4573]: I1203 09:58:52.338583 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc_fa5259c8-c545-4b45-a9c6-e0bca4d3aba9/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:52 crc kubenswrapper[4573]: I1203 09:58:52.395633 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b/tempest-tests-tempest-tests-runner/0.log" Dec 03 09:58:52 crc kubenswrapper[4573]: I1203 09:58:52.613556 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f5b253a6-9fcb-432b-81d9-6a6b980367c7/test-operator-logs-container/0.log" Dec 03 09:58:52 crc kubenswrapper[4573]: I1203 09:58:52.826011 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf_8ebf9479-f67b-443c-8ae2-3bec8e719750/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 09:58:55 crc kubenswrapper[4573]: I1203 09:58:55.030088 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:58:55 crc kubenswrapper[4573]: E1203 09:58:55.031607 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 09:59:09 crc kubenswrapper[4573]: I1203 09:59:09.030356 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 09:59:10 crc kubenswrapper[4573]: I1203 09:59:10.516710 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"ed8849625f09788d6b1f0caa2c569eee66c1bad7cd396148adfae49d1b223065"} Dec 03 09:59:11 crc kubenswrapper[4573]: I1203 09:59:11.231420 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_a7d45736-034b-4537-ba45-dd60f3bc3e7a/memcached/0.log" Dec 03 09:59:29 crc kubenswrapper[4573]: I1203 09:59:29.407913 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/util/0.log" Dec 03 09:59:29 crc kubenswrapper[4573]: I1203 09:59:29.653588 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/pull/0.log" Dec 03 09:59:29 crc kubenswrapper[4573]: I1203 09:59:29.713618 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/pull/0.log" Dec 03 09:59:29 crc kubenswrapper[4573]: I1203 09:59:29.762506 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/util/0.log" Dec 03 09:59:30 crc kubenswrapper[4573]: I1203 09:59:30.097135 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/util/0.log" Dec 03 09:59:30 crc kubenswrapper[4573]: I1203 09:59:30.097715 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/extract/0.log" Dec 03 09:59:30 crc kubenswrapper[4573]: I1203 09:59:30.155519 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/pull/0.log" Dec 03 09:59:30 crc kubenswrapper[4573]: I1203 09:59:30.378133 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-nx7mq_e27aea95-097f-4aa8-ae2b-7d212b8640b4/kube-rbac-proxy/0.log" Dec 03 09:59:30 crc kubenswrapper[4573]: I1203 09:59:30.511223 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-hb6l4_e0ae572b-e68c-48a0-8649-7eea884e61b0/kube-rbac-proxy/0.log" Dec 03 09:59:30 crc kubenswrapper[4573]: I1203 09:59:30.535011 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-nx7mq_e27aea95-097f-4aa8-ae2b-7d212b8640b4/manager/0.log" Dec 03 09:59:30 crc kubenswrapper[4573]: I1203 09:59:30.671597 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-hb6l4_e0ae572b-e68c-48a0-8649-7eea884e61b0/manager/0.log" Dec 03 09:59:31 crc kubenswrapper[4573]: I1203 09:59:31.358520 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-vl9xb_ee80479f-97a6-43ef-a026-d5bf1931d962/manager/0.log" Dec 03 09:59:31 crc kubenswrapper[4573]: I1203 09:59:31.403321 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-vl9xb_ee80479f-97a6-43ef-a026-d5bf1931d962/kube-rbac-proxy/0.log" Dec 03 09:59:31 crc kubenswrapper[4573]: I1203 09:59:31.517450 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-s8679_a855d891-6a41-477d-98d7-fcbba59aee28/kube-rbac-proxy/0.log" Dec 03 09:59:31 crc kubenswrapper[4573]: I1203 09:59:31.626752 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gnlgf_ff1aaf12-2fcd-41d0-a575-a0e5447fc11f/kube-rbac-proxy/0.log" Dec 03 09:59:31 crc kubenswrapper[4573]: I1203 09:59:31.698135 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-s8679_a855d891-6a41-477d-98d7-fcbba59aee28/manager/0.log" Dec 03 09:59:31 crc kubenswrapper[4573]: I1203 09:59:31.789107 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gnlgf_ff1aaf12-2fcd-41d0-a575-a0e5447fc11f/manager/0.log" Dec 03 09:59:31 crc kubenswrapper[4573]: I1203 09:59:31.861882 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2qnlp_622de9fb-c4ba-4727-8c22-5fcb36f39751/kube-rbac-proxy/0.log" Dec 03 09:59:31 crc kubenswrapper[4573]: I1203 09:59:31.977431 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2qnlp_622de9fb-c4ba-4727-8c22-5fcb36f39751/manager/0.log" Dec 03 09:59:32 crc kubenswrapper[4573]: I1203 09:59:32.108457 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gljjt_22bca7e1-22a7-4ee7-852b-25fddeb8fce3/kube-rbac-proxy/0.log" Dec 03 09:59:32 crc kubenswrapper[4573]: I1203 09:59:32.299427 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gljjt_22bca7e1-22a7-4ee7-852b-25fddeb8fce3/manager/0.log" Dec 03 09:59:32 crc kubenswrapper[4573]: I1203 09:59:32.352890 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-wcbtn_9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b/kube-rbac-proxy/0.log" Dec 03 09:59:32 crc kubenswrapper[4573]: I1203 09:59:32.410689 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-wcbtn_9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b/manager/0.log" Dec 03 09:59:32 crc kubenswrapper[4573]: I1203 09:59:32.710379 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-hdhrp_5aba2f13-92bd-4ea3-bfcb-7646909db04b/manager/0.log" Dec 03 09:59:32 crc kubenswrapper[4573]: I1203 09:59:32.746162 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-hdhrp_5aba2f13-92bd-4ea3-bfcb-7646909db04b/kube-rbac-proxy/0.log" Dec 03 09:59:32 crc kubenswrapper[4573]: I1203 09:59:32.927932 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qphvx_4613c040-3ca1-458c-8ff8-2b8858a7ad35/kube-rbac-proxy/0.log" Dec 03 09:59:32 crc kubenswrapper[4573]: I1203 09:59:32.987261 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qphvx_4613c040-3ca1-458c-8ff8-2b8858a7ad35/manager/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.072641 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hl72w_a2888fa7-1f19-48c8-853e-a3952cc93ec6/kube-rbac-proxy/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.262771 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hl72w_a2888fa7-1f19-48c8-853e-a3952cc93ec6/manager/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.293869 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-r6qvb_9d6a307e-5cf2-4a23-921c-c5b562494cb0/kube-rbac-proxy/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.342283 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-r6qvb_9d6a307e-5cf2-4a23-921c-c5b562494cb0/manager/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.579765 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-sdrbz_fd426f33-5f03-4fec-bde4-e72e9c762762/kube-rbac-proxy/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.627962 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-sdrbz_fd426f33-5f03-4fec-bde4-e72e9c762762/manager/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.678416 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-bnztb_98fcc67c-249b-4700-8434-c8dd4e1646ae/kube-rbac-proxy/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.754112 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-bnztb_98fcc67c-249b-4700-8434-c8dd4e1646ae/manager/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.913821 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg_a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629/kube-rbac-proxy/0.log" Dec 03 09:59:33 crc kubenswrapper[4573]: I1203 09:59:33.916152 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg_a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629/manager/0.log" Dec 03 09:59:34 crc kubenswrapper[4573]: I1203 09:59:34.249531 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-h4tdp_1b63239f-176e-4a3b-ac7d-605be09bd269/registry-server/0.log" Dec 03 09:59:34 crc kubenswrapper[4573]: I1203 09:59:34.460665 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-f6f67fdf8-vgsnr_42fe335f-c3e5-4792-a174-2f5aeaa7d360/operator/0.log" Dec 03 09:59:34 crc kubenswrapper[4573]: I1203 09:59:34.509995 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-d77h9_fd5e6711-117e-42b2-90e6-cfddf410315d/kube-rbac-proxy/0.log" Dec 03 09:59:34 crc kubenswrapper[4573]: I1203 09:59:34.798793 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-d77h9_fd5e6711-117e-42b2-90e6-cfddf410315d/manager/0.log" Dec 03 09:59:34 crc kubenswrapper[4573]: I1203 09:59:34.882261 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c5vc7_e94079e6-8701-48bf-ab58-1867fdf6e46e/manager/0.log" Dec 03 09:59:34 crc kubenswrapper[4573]: I1203 09:59:34.891628 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c5vc7_e94079e6-8701-48bf-ab58-1867fdf6e46e/kube-rbac-proxy/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.207229 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jbwlq_b4436129-0e91-443b-8a97-61fa7f8cc2e4/operator/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.265747 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-nbfxj_57f767e9-7c3a-4707-9b76-ed715c297238/kube-rbac-proxy/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.352903 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-nbfxj_57f767e9-7c3a-4707-9b76-ed715c297238/manager/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.476940 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-58f6d66c48-gwplc_5acea32c-fc74-40d8-b363-c64470ee880e/manager/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.493425 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-jd6bv_8f65995e-e70a-48e4-9504-8aa85d43ea9e/kube-rbac-proxy/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.593510 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-jd6bv_8f65995e-e70a-48e4-9504-8aa85d43ea9e/manager/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.728388 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wdvlj_6535c22e-f476-4067-b1ab-7e0ed5c23360/kube-rbac-proxy/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.735143 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wdvlj_6535c22e-f476-4067-b1ab-7e0ed5c23360/manager/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.797347 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-l5vf4_e6f9febf-6db7-4ea8-8b96-7a939102be2a/kube-rbac-proxy/0.log" Dec 03 09:59:35 crc kubenswrapper[4573]: I1203 09:59:35.867773 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-l5vf4_e6f9febf-6db7-4ea8-8b96-7a939102be2a/manager/0.log" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.564879 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hxvd5"] Dec 03 09:59:44 crc kubenswrapper[4573]: E1203 09:59:44.566369 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerName="registry-server" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.566388 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerName="registry-server" Dec 03 09:59:44 crc kubenswrapper[4573]: E1203 09:59:44.566422 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerName="extract-content" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.566427 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerName="extract-content" Dec 03 09:59:44 crc kubenswrapper[4573]: E1203 09:59:44.566460 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerName="extract-utilities" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.566467 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerName="extract-utilities" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.566708 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed85f45-6939-4da3-b93d-29d48d3dc433" containerName="registry-server" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.568608 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.603365 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hxvd5"] Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.644862 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6zmw\" (UniqueName: \"kubernetes.io/projected/657bcc0c-5649-4345-bf7c-b01c4ab82d64-kube-api-access-m6zmw\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.644972 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-utilities\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.645099 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-catalog-content\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.749370 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-utilities\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.750406 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-utilities\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.750540 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-catalog-content\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.750674 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6zmw\" (UniqueName: \"kubernetes.io/projected/657bcc0c-5649-4345-bf7c-b01c4ab82d64-kube-api-access-m6zmw\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.753164 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-catalog-content\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.775832 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6zmw\" (UniqueName: \"kubernetes.io/projected/657bcc0c-5649-4345-bf7c-b01c4ab82d64-kube-api-access-m6zmw\") pod \"redhat-operators-hxvd5\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:44 crc kubenswrapper[4573]: I1203 09:59:44.897308 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:45 crc kubenswrapper[4573]: I1203 09:59:45.537488 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hxvd5"] Dec 03 09:59:45 crc kubenswrapper[4573]: I1203 09:59:45.871282 4573 generic.go:334] "Generic (PLEG): container finished" podID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerID="87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb" exitCode=0 Dec 03 09:59:45 crc kubenswrapper[4573]: I1203 09:59:45.871660 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxvd5" event={"ID":"657bcc0c-5649-4345-bf7c-b01c4ab82d64","Type":"ContainerDied","Data":"87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb"} Dec 03 09:59:45 crc kubenswrapper[4573]: I1203 09:59:45.871841 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxvd5" event={"ID":"657bcc0c-5649-4345-bf7c-b01c4ab82d64","Type":"ContainerStarted","Data":"2e03f4ed5999c7ca92207aa829f35495c0aa6aea6d15837fa7d883ca40393bc5"} Dec 03 09:59:46 crc kubenswrapper[4573]: I1203 09:59:46.886314 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxvd5" event={"ID":"657bcc0c-5649-4345-bf7c-b01c4ab82d64","Type":"ContainerStarted","Data":"9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2"} Dec 03 09:59:50 crc kubenswrapper[4573]: I1203 09:59:50.926249 4573 generic.go:334] "Generic (PLEG): container finished" podID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerID="9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2" exitCode=0 Dec 03 09:59:50 crc kubenswrapper[4573]: I1203 09:59:50.926339 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxvd5" event={"ID":"657bcc0c-5649-4345-bf7c-b01c4ab82d64","Type":"ContainerDied","Data":"9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2"} Dec 03 09:59:51 crc kubenswrapper[4573]: I1203 09:59:51.939008 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxvd5" event={"ID":"657bcc0c-5649-4345-bf7c-b01c4ab82d64","Type":"ContainerStarted","Data":"e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993"} Dec 03 09:59:51 crc kubenswrapper[4573]: I1203 09:59:51.959428 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hxvd5" podStartSLOduration=2.40722635 podStartE2EDuration="7.959381282s" podCreationTimestamp="2025-12-03 09:59:44 +0000 UTC" firstStartedPulling="2025-12-03 09:59:45.87463206 +0000 UTC m=+4906.443011319" lastFinishedPulling="2025-12-03 09:59:51.426786992 +0000 UTC m=+4911.995166251" observedRunningTime="2025-12-03 09:59:51.956404091 +0000 UTC m=+4912.524783360" watchObservedRunningTime="2025-12-03 09:59:51.959381282 +0000 UTC m=+4912.527760541" Dec 03 09:59:54 crc kubenswrapper[4573]: I1203 09:59:54.897532 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:54 crc kubenswrapper[4573]: I1203 09:59:54.897892 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 09:59:55 crc kubenswrapper[4573]: I1203 09:59:55.944499 4573 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hxvd5" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="registry-server" probeResult="failure" output=< Dec 03 09:59:55 crc kubenswrapper[4573]: timeout: failed to connect service ":50051" within 1s Dec 03 09:59:55 crc kubenswrapper[4573]: > Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.155776 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws"] Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.158028 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.161421 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.161421 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.170683 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws"] Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.268298 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-config-volume\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.268369 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx8d8\" (UniqueName: \"kubernetes.io/projected/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-kube-api-access-dx8d8\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.268745 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-secret-volume\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.370805 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-secret-volume\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.370994 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-config-volume\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.371037 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx8d8\" (UniqueName: \"kubernetes.io/projected/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-kube-api-access-dx8d8\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.372444 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-config-volume\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.393965 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-secret-volume\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.394167 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx8d8\" (UniqueName: \"kubernetes.io/projected/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-kube-api-access-dx8d8\") pod \"collect-profiles-29412600-g6dws\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:00 crc kubenswrapper[4573]: I1203 10:00:00.484870 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:01 crc kubenswrapper[4573]: I1203 10:00:01.007936 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws"] Dec 03 10:00:01 crc kubenswrapper[4573]: W1203 10:00:01.014005 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeabf5e75_6bf4_441d_9fd5_8b5c94f9f537.slice/crio-f1b1d2e35715dc1c6009a38817206d3a0bbc974ecb35441cd91458896b72a31b WatchSource:0}: Error finding container f1b1d2e35715dc1c6009a38817206d3a0bbc974ecb35441cd91458896b72a31b: Status 404 returned error can't find the container with id f1b1d2e35715dc1c6009a38817206d3a0bbc974ecb35441cd91458896b72a31b Dec 03 10:00:02 crc kubenswrapper[4573]: I1203 10:00:02.032712 4573 generic.go:334] "Generic (PLEG): container finished" podID="eabf5e75-6bf4-441d-9fd5-8b5c94f9f537" containerID="2075b95a9bb8e8548b9cb3c2ab1bb8e5fd92892798ebcf0358d5eb5850bb960e" exitCode=0 Dec 03 10:00:02 crc kubenswrapper[4573]: I1203 10:00:02.049116 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" event={"ID":"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537","Type":"ContainerDied","Data":"2075b95a9bb8e8548b9cb3c2ab1bb8e5fd92892798ebcf0358d5eb5850bb960e"} Dec 03 10:00:02 crc kubenswrapper[4573]: I1203 10:00:02.049186 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" event={"ID":"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537","Type":"ContainerStarted","Data":"f1b1d2e35715dc1c6009a38817206d3a0bbc974ecb35441cd91458896b72a31b"} Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.570440 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.750945 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-config-volume\") pod \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.751098 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-secret-volume\") pod \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.751306 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx8d8\" (UniqueName: \"kubernetes.io/projected/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-kube-api-access-dx8d8\") pod \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\" (UID: \"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537\") " Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.751615 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-config-volume" (OuterVolumeSpecName: "config-volume") pod "eabf5e75-6bf4-441d-9fd5-8b5c94f9f537" (UID: "eabf5e75-6bf4-441d-9fd5-8b5c94f9f537"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.751923 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.765410 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eabf5e75-6bf4-441d-9fd5-8b5c94f9f537" (UID: "eabf5e75-6bf4-441d-9fd5-8b5c94f9f537"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.765663 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-kube-api-access-dx8d8" (OuterVolumeSpecName: "kube-api-access-dx8d8") pod "eabf5e75-6bf4-441d-9fd5-8b5c94f9f537" (UID: "eabf5e75-6bf4-441d-9fd5-8b5c94f9f537"). InnerVolumeSpecName "kube-api-access-dx8d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.853400 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx8d8\" (UniqueName: \"kubernetes.io/projected/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-kube-api-access-dx8d8\") on node \"crc\" DevicePath \"\"" Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.853765 4573 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eabf5e75-6bf4-441d-9fd5-8b5c94f9f537-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 10:00:03 crc kubenswrapper[4573]: I1203 10:00:03.945992 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-7mk94_192416c5-8025-4dd5-840b-9954ce7d009b/control-plane-machine-set-operator/0.log" Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.064935 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" event={"ID":"eabf5e75-6bf4-441d-9fd5-8b5c94f9f537","Type":"ContainerDied","Data":"f1b1d2e35715dc1c6009a38817206d3a0bbc974ecb35441cd91458896b72a31b"} Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.064983 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412600-g6dws" Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.064991 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1b1d2e35715dc1c6009a38817206d3a0bbc974ecb35441cd91458896b72a31b" Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.187605 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8qfqs_284d9db2-04e5-4c39-b4cf-6f001de356c4/kube-rbac-proxy/0.log" Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.239895 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8qfqs_284d9db2-04e5-4c39-b4cf-6f001de356c4/machine-api-operator/0.log" Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.704027 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn"] Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.716842 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412555-z8gxn"] Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.944830 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 10:00:04 crc kubenswrapper[4573]: I1203 10:00:04.992606 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 10:00:05 crc kubenswrapper[4573]: I1203 10:00:05.187373 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hxvd5"] Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.065547 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dc0c400-a678-4df2-9e5b-f7163404f664" path="/var/lib/kubelet/pods/4dc0c400-a678-4df2-9e5b-f7163404f664/volumes" Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.085887 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hxvd5" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="registry-server" containerID="cri-o://e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993" gracePeriod=2 Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.626859 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.737635 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-catalog-content\") pod \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.737836 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6zmw\" (UniqueName: \"kubernetes.io/projected/657bcc0c-5649-4345-bf7c-b01c4ab82d64-kube-api-access-m6zmw\") pod \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.738889 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-utilities\") pod \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\" (UID: \"657bcc0c-5649-4345-bf7c-b01c4ab82d64\") " Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.739686 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-utilities" (OuterVolumeSpecName: "utilities") pod "657bcc0c-5649-4345-bf7c-b01c4ab82d64" (UID: "657bcc0c-5649-4345-bf7c-b01c4ab82d64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.757439 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/657bcc0c-5649-4345-bf7c-b01c4ab82d64-kube-api-access-m6zmw" (OuterVolumeSpecName: "kube-api-access-m6zmw") pod "657bcc0c-5649-4345-bf7c-b01c4ab82d64" (UID: "657bcc0c-5649-4345-bf7c-b01c4ab82d64"). InnerVolumeSpecName "kube-api-access-m6zmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.840411 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.840457 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6zmw\" (UniqueName: \"kubernetes.io/projected/657bcc0c-5649-4345-bf7c-b01c4ab82d64-kube-api-access-m6zmw\") on node \"crc\" DevicePath \"\"" Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.854440 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "657bcc0c-5649-4345-bf7c-b01c4ab82d64" (UID: "657bcc0c-5649-4345-bf7c-b01c4ab82d64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:00:06 crc kubenswrapper[4573]: I1203 10:00:06.942423 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/657bcc0c-5649-4345-bf7c-b01c4ab82d64-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.096462 4573 generic.go:334] "Generic (PLEG): container finished" podID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerID="e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993" exitCode=0 Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.096504 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxvd5" event={"ID":"657bcc0c-5649-4345-bf7c-b01c4ab82d64","Type":"ContainerDied","Data":"e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993"} Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.096567 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hxvd5" event={"ID":"657bcc0c-5649-4345-bf7c-b01c4ab82d64","Type":"ContainerDied","Data":"2e03f4ed5999c7ca92207aa829f35495c0aa6aea6d15837fa7d883ca40393bc5"} Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.096611 4573 scope.go:117] "RemoveContainer" containerID="e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.097816 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hxvd5" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.134014 4573 scope.go:117] "RemoveContainer" containerID="9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.154801 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hxvd5"] Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.161618 4573 scope.go:117] "RemoveContainer" containerID="87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.172791 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hxvd5"] Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.229029 4573 scope.go:117] "RemoveContainer" containerID="e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993" Dec 03 10:00:07 crc kubenswrapper[4573]: E1203 10:00:07.229634 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993\": container with ID starting with e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993 not found: ID does not exist" containerID="e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.229692 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993"} err="failed to get container status \"e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993\": rpc error: code = NotFound desc = could not find container \"e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993\": container with ID starting with e6cfb7a595f2d0e10cb1d2db8b7889ecb4519a2bb0e564b8167559ec14334993 not found: ID does not exist" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.229728 4573 scope.go:117] "RemoveContainer" containerID="9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2" Dec 03 10:00:07 crc kubenswrapper[4573]: E1203 10:00:07.230579 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2\": container with ID starting with 9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2 not found: ID does not exist" containerID="9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.230652 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2"} err="failed to get container status \"9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2\": rpc error: code = NotFound desc = could not find container \"9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2\": container with ID starting with 9062065ea027eee5e6a0428a6b4ac8bb114bb4ad811f10b6b18c303baa0dbfc2 not found: ID does not exist" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.230690 4573 scope.go:117] "RemoveContainer" containerID="87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb" Dec 03 10:00:07 crc kubenswrapper[4573]: E1203 10:00:07.231062 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb\": container with ID starting with 87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb not found: ID does not exist" containerID="87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb" Dec 03 10:00:07 crc kubenswrapper[4573]: I1203 10:00:07.231098 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb"} err="failed to get container status \"87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb\": rpc error: code = NotFound desc = could not find container \"87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb\": container with ID starting with 87e44bcb48d134ec8d54dc850acee5b7e2a88add462bc606c5bb22ee7ad4d5fb not found: ID does not exist" Dec 03 10:00:08 crc kubenswrapper[4573]: I1203 10:00:08.047815 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" path="/var/lib/kubelet/pods/657bcc0c-5649-4345-bf7c-b01c4ab82d64/volumes" Dec 03 10:00:09 crc kubenswrapper[4573]: I1203 10:00:09.676192 4573 scope.go:117] "RemoveContainer" containerID="1fabd5357e3df26b564ad7aaea88f5939ff8ac24eec6d7e65dfcc406ddb2cf7b" Dec 03 10:00:19 crc kubenswrapper[4573]: I1203 10:00:19.703567 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-h9bp4_ee0f7af7-48fb-433a-8f6b-801fd53f331d/cert-manager-controller/0.log" Dec 03 10:00:19 crc kubenswrapper[4573]: I1203 10:00:19.873268 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-4226f_9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf/cert-manager-cainjector/0.log" Dec 03 10:00:20 crc kubenswrapper[4573]: I1203 10:00:20.009173 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-twvqd_f4edafce-1ec1-4ada-9861-c36ce8c3906f/cert-manager-webhook/0.log" Dec 03 10:00:35 crc kubenswrapper[4573]: I1203 10:00:35.161799 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-9wr28_7006af64-f3a2-4410-84e4-f6dfd3bcf0ae/nmstate-console-plugin/0.log" Dec 03 10:00:35 crc kubenswrapper[4573]: I1203 10:00:35.419716 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6mwnn_72904583-c577-4d6a-91e1-3ec1d51ec7c2/nmstate-handler/0.log" Dec 03 10:00:35 crc kubenswrapper[4573]: I1203 10:00:35.473840 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-g2mdj_483c74ea-46d0-42e4-84c3-35ed45a0507e/kube-rbac-proxy/0.log" Dec 03 10:00:35 crc kubenswrapper[4573]: I1203 10:00:35.515689 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-g2mdj_483c74ea-46d0-42e4-84c3-35ed45a0507e/nmstate-metrics/0.log" Dec 03 10:00:35 crc kubenswrapper[4573]: I1203 10:00:35.683409 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-j7vz4_d8f81a47-9022-45cf-93e5-8989ee9121ce/nmstate-operator/0.log" Dec 03 10:00:35 crc kubenswrapper[4573]: I1203 10:00:35.798433 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-hlbnj_358ba50e-2a33-45a2-b1d4-b606638246d5/nmstate-webhook/0.log" Dec 03 10:00:54 crc kubenswrapper[4573]: I1203 10:00:54.828658 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rj7p6_974ec28d-5851-4dd3-b992-9920ed8d32de/kube-rbac-proxy/0.log" Dec 03 10:00:54 crc kubenswrapper[4573]: I1203 10:00:54.996713 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rj7p6_974ec28d-5851-4dd3-b992-9920ed8d32de/controller/0.log" Dec 03 10:00:55 crc kubenswrapper[4573]: I1203 10:00:55.090143 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-frr-files/0.log" Dec 03 10:00:55 crc kubenswrapper[4573]: I1203 10:00:55.379634 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-frr-files/0.log" Dec 03 10:00:55 crc kubenswrapper[4573]: I1203 10:00:55.405003 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-metrics/0.log" Dec 03 10:00:55 crc kubenswrapper[4573]: I1203 10:00:55.461729 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-reloader/0.log" Dec 03 10:00:55 crc kubenswrapper[4573]: I1203 10:00:55.471864 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-reloader/0.log" Dec 03 10:00:55 crc kubenswrapper[4573]: I1203 10:00:55.949871 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-reloader/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.018127 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-frr-files/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.025190 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-metrics/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.096631 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-metrics/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.258686 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-reloader/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.294230 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-frr-files/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.299115 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-metrics/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.403724 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/controller/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.546706 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/frr-metrics/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.632451 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/kube-rbac-proxy/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.826582 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/kube-rbac-proxy-frr/0.log" Dec 03 10:00:56 crc kubenswrapper[4573]: I1203 10:00:56.872413 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/reloader/0.log" Dec 03 10:00:57 crc kubenswrapper[4573]: I1203 10:00:57.113868 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-cngn6_0e088580-c78b-42cc-b013-bc58bcffa838/frr-k8s-webhook-server/0.log" Dec 03 10:00:57 crc kubenswrapper[4573]: I1203 10:00:57.369503 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-d75bb676c-wwfz6_29c77033-f8ba-44b5-a275-0861c60318bf/manager/0.log" Dec 03 10:00:57 crc kubenswrapper[4573]: I1203 10:00:57.486834 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-b9cfd44fd-mswlh_4e051555-a35f-45f0-ac65-237da26aa296/webhook-server/0.log" Dec 03 10:00:57 crc kubenswrapper[4573]: I1203 10:00:57.796784 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/frr/0.log" Dec 03 10:00:57 crc kubenswrapper[4573]: I1203 10:00:57.902233 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hw9mt_09e37261-797a-4fb1-9ae2-5de662b865e8/kube-rbac-proxy/0.log" Dec 03 10:00:58 crc kubenswrapper[4573]: I1203 10:00:58.157283 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hw9mt_09e37261-797a-4fb1-9ae2-5de662b865e8/speaker/0.log" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.187955 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29412601-xpbj9"] Dec 03 10:01:00 crc kubenswrapper[4573]: E1203 10:01:00.188720 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="extract-utilities" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.188735 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="extract-utilities" Dec 03 10:01:00 crc kubenswrapper[4573]: E1203 10:01:00.188775 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="registry-server" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.188787 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="registry-server" Dec 03 10:01:00 crc kubenswrapper[4573]: E1203 10:01:00.188797 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eabf5e75-6bf4-441d-9fd5-8b5c94f9f537" containerName="collect-profiles" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.188804 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="eabf5e75-6bf4-441d-9fd5-8b5c94f9f537" containerName="collect-profiles" Dec 03 10:01:00 crc kubenswrapper[4573]: E1203 10:01:00.188821 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="extract-content" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.188828 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="extract-content" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.189012 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="eabf5e75-6bf4-441d-9fd5-8b5c94f9f537" containerName="collect-profiles" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.189068 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="657bcc0c-5649-4345-bf7c-b01c4ab82d64" containerName="registry-server" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.189840 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.210091 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412601-xpbj9"] Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.309814 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-fernet-keys\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.309873 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-combined-ca-bundle\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.309904 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-config-data\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.309953 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nxn8\" (UniqueName: \"kubernetes.io/projected/928d502e-8067-4583-a118-206953580db2-kube-api-access-4nxn8\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.411833 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-fernet-keys\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.411892 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-combined-ca-bundle\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.411937 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-config-data\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.411997 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nxn8\" (UniqueName: \"kubernetes.io/projected/928d502e-8067-4583-a118-206953580db2-kube-api-access-4nxn8\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.431847 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nxn8\" (UniqueName: \"kubernetes.io/projected/928d502e-8067-4583-a118-206953580db2-kube-api-access-4nxn8\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.432091 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-combined-ca-bundle\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.432760 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-fernet-keys\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.435519 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-config-data\") pod \"keystone-cron-29412601-xpbj9\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:00 crc kubenswrapper[4573]: I1203 10:01:00.520914 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:01 crc kubenswrapper[4573]: I1203 10:01:01.023479 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412601-xpbj9"] Dec 03 10:01:01 crc kubenswrapper[4573]: I1203 10:01:01.628528 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412601-xpbj9" event={"ID":"928d502e-8067-4583-a118-206953580db2","Type":"ContainerStarted","Data":"ec87e592128413a10c5679d40d2c7151aaff41818e69be67a44e9a69777ee8c9"} Dec 03 10:01:01 crc kubenswrapper[4573]: I1203 10:01:01.628805 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412601-xpbj9" event={"ID":"928d502e-8067-4583-a118-206953580db2","Type":"ContainerStarted","Data":"69c945c3be9d086b14104c80b03c1b515a628e40fd8507efc696ebe5d6912ca2"} Dec 03 10:01:01 crc kubenswrapper[4573]: I1203 10:01:01.662963 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29412601-xpbj9" podStartSLOduration=1.662945488 podStartE2EDuration="1.662945488s" podCreationTimestamp="2025-12-03 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:01:01.662070424 +0000 UTC m=+4982.230449693" watchObservedRunningTime="2025-12-03 10:01:01.662945488 +0000 UTC m=+4982.231324747" Dec 03 10:01:04 crc kubenswrapper[4573]: I1203 10:01:04.660721 4573 generic.go:334] "Generic (PLEG): container finished" podID="928d502e-8067-4583-a118-206953580db2" containerID="ec87e592128413a10c5679d40d2c7151aaff41818e69be67a44e9a69777ee8c9" exitCode=0 Dec 03 10:01:04 crc kubenswrapper[4573]: I1203 10:01:04.660796 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412601-xpbj9" event={"ID":"928d502e-8067-4583-a118-206953580db2","Type":"ContainerDied","Data":"ec87e592128413a10c5679d40d2c7151aaff41818e69be67a44e9a69777ee8c9"} Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.023158 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.219018 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nxn8\" (UniqueName: \"kubernetes.io/projected/928d502e-8067-4583-a118-206953580db2-kube-api-access-4nxn8\") pod \"928d502e-8067-4583-a118-206953580db2\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.219190 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-combined-ca-bundle\") pod \"928d502e-8067-4583-a118-206953580db2\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.219350 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-fernet-keys\") pod \"928d502e-8067-4583-a118-206953580db2\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.219409 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-config-data\") pod \"928d502e-8067-4583-a118-206953580db2\" (UID: \"928d502e-8067-4583-a118-206953580db2\") " Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.527291 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "928d502e-8067-4583-a118-206953580db2" (UID: "928d502e-8067-4583-a118-206953580db2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.527363 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/928d502e-8067-4583-a118-206953580db2-kube-api-access-4nxn8" (OuterVolumeSpecName: "kube-api-access-4nxn8") pod "928d502e-8067-4583-a118-206953580db2" (UID: "928d502e-8067-4583-a118-206953580db2"). InnerVolumeSpecName "kube-api-access-4nxn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.627515 4573 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.627558 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nxn8\" (UniqueName: \"kubernetes.io/projected/928d502e-8067-4583-a118-206953580db2-kube-api-access-4nxn8\") on node \"crc\" DevicePath \"\"" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.645814 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "928d502e-8067-4583-a118-206953580db2" (UID: "928d502e-8067-4583-a118-206953580db2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.675826 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-config-data" (OuterVolumeSpecName: "config-data") pod "928d502e-8067-4583-a118-206953580db2" (UID: "928d502e-8067-4583-a118-206953580db2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.687261 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412601-xpbj9" event={"ID":"928d502e-8067-4583-a118-206953580db2","Type":"ContainerDied","Data":"69c945c3be9d086b14104c80b03c1b515a628e40fd8507efc696ebe5d6912ca2"} Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.687437 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69c945c3be9d086b14104c80b03c1b515a628e40fd8507efc696ebe5d6912ca2" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.687314 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412601-xpbj9" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.730416 4573 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 10:01:06 crc kubenswrapper[4573]: I1203 10:01:06.730472 4573 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928d502e-8067-4583-a118-206953580db2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 10:01:13 crc kubenswrapper[4573]: I1203 10:01:13.574652 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/util/0.log" Dec 03 10:01:13 crc kubenswrapper[4573]: I1203 10:01:13.763605 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/util/0.log" Dec 03 10:01:13 crc kubenswrapper[4573]: I1203 10:01:13.773320 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/pull/0.log" Dec 03 10:01:13 crc kubenswrapper[4573]: I1203 10:01:13.816104 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/pull/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.075329 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/pull/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.082625 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/extract/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.100366 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/util/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.299817 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/util/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.508446 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/pull/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.515363 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/util/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.536370 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/pull/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.765177 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/pull/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.803744 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/util/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.841664 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/extract/0.log" Dec 03 10:01:14 crc kubenswrapper[4573]: I1203 10:01:14.979294 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-utilities/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.156880 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-utilities/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.175957 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-content/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.181173 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-content/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.397030 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-utilities/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.397436 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-content/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.669844 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-utilities/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.713285 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/registry-server/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.899424 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-content/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.910096 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-utilities/0.log" Dec 03 10:01:15 crc kubenswrapper[4573]: I1203 10:01:15.980360 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-content/0.log" Dec 03 10:01:16 crc kubenswrapper[4573]: I1203 10:01:16.129816 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-content/0.log" Dec 03 10:01:16 crc kubenswrapper[4573]: I1203 10:01:16.171415 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-utilities/0.log" Dec 03 10:01:16 crc kubenswrapper[4573]: I1203 10:01:16.617274 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hgwmj_e82d702a-82de-47e9-89fe-8ca23ce8a98b/marketplace-operator/0.log" Dec 03 10:01:16 crc kubenswrapper[4573]: I1203 10:01:16.683646 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-utilities/0.log" Dec 03 10:01:16 crc kubenswrapper[4573]: I1203 10:01:16.995905 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-utilities/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.031983 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/registry-server/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.107833 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-content/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.139283 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-content/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.285547 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-content/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.336509 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-utilities/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.668036 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/registry-server/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.668429 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-utilities/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.871836 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-utilities/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.911665 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-content/0.log" Dec 03 10:01:17 crc kubenswrapper[4573]: I1203 10:01:17.949264 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-content/0.log" Dec 03 10:01:18 crc kubenswrapper[4573]: I1203 10:01:18.185337 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-utilities/0.log" Dec 03 10:01:18 crc kubenswrapper[4573]: I1203 10:01:18.208438 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-content/0.log" Dec 03 10:01:18 crc kubenswrapper[4573]: I1203 10:01:18.780417 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/registry-server/0.log" Dec 03 10:01:26 crc kubenswrapper[4573]: I1203 10:01:26.943808 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:01:26 crc kubenswrapper[4573]: I1203 10:01:26.944391 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:01:56 crc kubenswrapper[4573]: I1203 10:01:56.943297 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:01:56 crc kubenswrapper[4573]: I1203 10:01:56.944785 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:02:26 crc kubenswrapper[4573]: I1203 10:02:26.943523 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:02:26 crc kubenswrapper[4573]: I1203 10:02:26.944082 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:02:26 crc kubenswrapper[4573]: I1203 10:02:26.944136 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 10:02:26 crc kubenswrapper[4573]: I1203 10:02:26.944990 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed8849625f09788d6b1f0caa2c569eee66c1bad7cd396148adfae49d1b223065"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 10:02:26 crc kubenswrapper[4573]: I1203 10:02:26.945067 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://ed8849625f09788d6b1f0caa2c569eee66c1bad7cd396148adfae49d1b223065" gracePeriod=600 Dec 03 10:02:27 crc kubenswrapper[4573]: I1203 10:02:27.457975 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="ed8849625f09788d6b1f0caa2c569eee66c1bad7cd396148adfae49d1b223065" exitCode=0 Dec 03 10:02:27 crc kubenswrapper[4573]: I1203 10:02:27.458283 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"ed8849625f09788d6b1f0caa2c569eee66c1bad7cd396148adfae49d1b223065"} Dec 03 10:02:27 crc kubenswrapper[4573]: I1203 10:02:27.458429 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2"} Dec 03 10:02:27 crc kubenswrapper[4573]: I1203 10:02:27.458460 4573 scope.go:117] "RemoveContainer" containerID="f5328a88e4368d11fb9fab12bae0b27dfbe2e56bd96ea74942c2db4eddad04d7" Dec 03 10:03:02 crc kubenswrapper[4573]: I1203 10:03:02.969554 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-558rz"] Dec 03 10:03:02 crc kubenswrapper[4573]: E1203 10:03:02.971074 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="928d502e-8067-4583-a118-206953580db2" containerName="keystone-cron" Dec 03 10:03:02 crc kubenswrapper[4573]: I1203 10:03:02.971092 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="928d502e-8067-4583-a118-206953580db2" containerName="keystone-cron" Dec 03 10:03:02 crc kubenswrapper[4573]: I1203 10:03:02.971346 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="928d502e-8067-4583-a118-206953580db2" containerName="keystone-cron" Dec 03 10:03:02 crc kubenswrapper[4573]: I1203 10:03:02.973354 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:02 crc kubenswrapper[4573]: I1203 10:03:02.988080 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-558rz"] Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.005614 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fchg\" (UniqueName: \"kubernetes.io/projected/fecf37f5-06b8-4922-8136-a1453c5ed505-kube-api-access-7fchg\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.005698 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-catalog-content\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.005724 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-utilities\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.107172 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-catalog-content\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.107501 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-utilities\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.107719 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fchg\" (UniqueName: \"kubernetes.io/projected/fecf37f5-06b8-4922-8136-a1453c5ed505-kube-api-access-7fchg\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.108240 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-catalog-content\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.109153 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-utilities\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.129723 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fchg\" (UniqueName: \"kubernetes.io/projected/fecf37f5-06b8-4922-8136-a1453c5ed505-kube-api-access-7fchg\") pod \"community-operators-558rz\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.307453 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:03 crc kubenswrapper[4573]: I1203 10:03:03.948735 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-558rz"] Dec 03 10:03:03 crc kubenswrapper[4573]: W1203 10:03:03.965611 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfecf37f5_06b8_4922_8136_a1453c5ed505.slice/crio-198edd4fcd6199597061db9540f1464847db3a033990147fe6e7756c4f609156 WatchSource:0}: Error finding container 198edd4fcd6199597061db9540f1464847db3a033990147fe6e7756c4f609156: Status 404 returned error can't find the container with id 198edd4fcd6199597061db9540f1464847db3a033990147fe6e7756c4f609156 Dec 03 10:03:04 crc kubenswrapper[4573]: I1203 10:03:04.899129 4573 generic.go:334] "Generic (PLEG): container finished" podID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerID="ef09bc60621227376f05c80aa7a55c96f63c188fa1a4f491e9d1b10fcd18744b" exitCode=0 Dec 03 10:03:04 crc kubenswrapper[4573]: I1203 10:03:04.900269 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-558rz" event={"ID":"fecf37f5-06b8-4922-8136-a1453c5ed505","Type":"ContainerDied","Data":"ef09bc60621227376f05c80aa7a55c96f63c188fa1a4f491e9d1b10fcd18744b"} Dec 03 10:03:04 crc kubenswrapper[4573]: I1203 10:03:04.900358 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-558rz" event={"ID":"fecf37f5-06b8-4922-8136-a1453c5ed505","Type":"ContainerStarted","Data":"198edd4fcd6199597061db9540f1464847db3a033990147fe6e7756c4f609156"} Dec 03 10:03:05 crc kubenswrapper[4573]: I1203 10:03:05.916212 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-558rz" event={"ID":"fecf37f5-06b8-4922-8136-a1453c5ed505","Type":"ContainerStarted","Data":"ac1d0799691e094b2f275c12720396476dfb5d718e2f4d68705fb489a53e404f"} Dec 03 10:03:06 crc kubenswrapper[4573]: I1203 10:03:06.944793 4573 generic.go:334] "Generic (PLEG): container finished" podID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerID="ac1d0799691e094b2f275c12720396476dfb5d718e2f4d68705fb489a53e404f" exitCode=0 Dec 03 10:03:06 crc kubenswrapper[4573]: I1203 10:03:06.946411 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-558rz" event={"ID":"fecf37f5-06b8-4922-8136-a1453c5ed505","Type":"ContainerDied","Data":"ac1d0799691e094b2f275c12720396476dfb5d718e2f4d68705fb489a53e404f"} Dec 03 10:03:07 crc kubenswrapper[4573]: I1203 10:03:07.956066 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-558rz" event={"ID":"fecf37f5-06b8-4922-8136-a1453c5ed505","Type":"ContainerStarted","Data":"d136be8cf996261c502aa28a085c4ea745c22c570ba45415e3392f7fe209ca0a"} Dec 03 10:03:09 crc kubenswrapper[4573]: I1203 10:03:09.801844 4573 scope.go:117] "RemoveContainer" containerID="b7931a2ae1256a9b9465a3f3fd6bedf64e7d2e573778be469fb2c9b8d013a1d7" Dec 03 10:03:13 crc kubenswrapper[4573]: I1203 10:03:13.307742 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:13 crc kubenswrapper[4573]: I1203 10:03:13.312486 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:13 crc kubenswrapper[4573]: I1203 10:03:13.372382 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:13 crc kubenswrapper[4573]: I1203 10:03:13.402623 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-558rz" podStartSLOduration=8.793980165 podStartE2EDuration="11.402604176s" podCreationTimestamp="2025-12-03 10:03:02 +0000 UTC" firstStartedPulling="2025-12-03 10:03:04.905760423 +0000 UTC m=+5105.474139722" lastFinishedPulling="2025-12-03 10:03:07.514384474 +0000 UTC m=+5108.082763733" observedRunningTime="2025-12-03 10:03:07.978002576 +0000 UTC m=+5108.546381835" watchObservedRunningTime="2025-12-03 10:03:13.402604176 +0000 UTC m=+5113.970983435" Dec 03 10:03:14 crc kubenswrapper[4573]: I1203 10:03:14.064042 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:14 crc kubenswrapper[4573]: I1203 10:03:14.138908 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-558rz"] Dec 03 10:03:16 crc kubenswrapper[4573]: I1203 10:03:16.024570 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-558rz" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerName="registry-server" containerID="cri-o://d136be8cf996261c502aa28a085c4ea745c22c570ba45415e3392f7fe209ca0a" gracePeriod=2 Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.046348 4573 generic.go:334] "Generic (PLEG): container finished" podID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerID="d136be8cf996261c502aa28a085c4ea745c22c570ba45415e3392f7fe209ca0a" exitCode=0 Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.046442 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-558rz" event={"ID":"fecf37f5-06b8-4922-8136-a1453c5ed505","Type":"ContainerDied","Data":"d136be8cf996261c502aa28a085c4ea745c22c570ba45415e3392f7fe209ca0a"} Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.146182 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.275518 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-utilities\") pod \"fecf37f5-06b8-4922-8136-a1453c5ed505\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.276032 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-catalog-content\") pod \"fecf37f5-06b8-4922-8136-a1453c5ed505\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.276268 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fchg\" (UniqueName: \"kubernetes.io/projected/fecf37f5-06b8-4922-8136-a1453c5ed505-kube-api-access-7fchg\") pod \"fecf37f5-06b8-4922-8136-a1453c5ed505\" (UID: \"fecf37f5-06b8-4922-8136-a1453c5ed505\") " Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.276476 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-utilities" (OuterVolumeSpecName: "utilities") pod "fecf37f5-06b8-4922-8136-a1453c5ed505" (UID: "fecf37f5-06b8-4922-8136-a1453c5ed505"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.276999 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.285355 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fecf37f5-06b8-4922-8136-a1453c5ed505-kube-api-access-7fchg" (OuterVolumeSpecName: "kube-api-access-7fchg") pod "fecf37f5-06b8-4922-8136-a1453c5ed505" (UID: "fecf37f5-06b8-4922-8136-a1453c5ed505"). InnerVolumeSpecName "kube-api-access-7fchg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.346200 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fecf37f5-06b8-4922-8136-a1453c5ed505" (UID: "fecf37f5-06b8-4922-8136-a1453c5ed505"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.379357 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fecf37f5-06b8-4922-8136-a1453c5ed505-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:03:17 crc kubenswrapper[4573]: I1203 10:03:17.379626 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fchg\" (UniqueName: \"kubernetes.io/projected/fecf37f5-06b8-4922-8136-a1453c5ed505-kube-api-access-7fchg\") on node \"crc\" DevicePath \"\"" Dec 03 10:03:18 crc kubenswrapper[4573]: I1203 10:03:18.058388 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-558rz" event={"ID":"fecf37f5-06b8-4922-8136-a1453c5ed505","Type":"ContainerDied","Data":"198edd4fcd6199597061db9540f1464847db3a033990147fe6e7756c4f609156"} Dec 03 10:03:18 crc kubenswrapper[4573]: I1203 10:03:18.058467 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-558rz" Dec 03 10:03:18 crc kubenswrapper[4573]: I1203 10:03:18.058763 4573 scope.go:117] "RemoveContainer" containerID="d136be8cf996261c502aa28a085c4ea745c22c570ba45415e3392f7fe209ca0a" Dec 03 10:03:18 crc kubenswrapper[4573]: I1203 10:03:18.085505 4573 scope.go:117] "RemoveContainer" containerID="ac1d0799691e094b2f275c12720396476dfb5d718e2f4d68705fb489a53e404f" Dec 03 10:03:18 crc kubenswrapper[4573]: I1203 10:03:18.095607 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-558rz"] Dec 03 10:03:18 crc kubenswrapper[4573]: I1203 10:03:18.104341 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-558rz"] Dec 03 10:03:18 crc kubenswrapper[4573]: I1203 10:03:18.655539 4573 scope.go:117] "RemoveContainer" containerID="ef09bc60621227376f05c80aa7a55c96f63c188fa1a4f491e9d1b10fcd18744b" Dec 03 10:03:20 crc kubenswrapper[4573]: I1203 10:03:20.042793 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" path="/var/lib/kubelet/pods/fecf37f5-06b8-4922-8136-a1453c5ed505/volumes" Dec 03 10:03:50 crc kubenswrapper[4573]: I1203 10:03:50.419749 4573 generic.go:334] "Generic (PLEG): container finished" podID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerID="1cf28177f9adbcf0a9c79a12471b1a52b0cd46040c717fee0d7d425a0e157447" exitCode=0 Dec 03 10:03:50 crc kubenswrapper[4573]: I1203 10:03:50.419826 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vzks7/must-gather-658jg" event={"ID":"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f","Type":"ContainerDied","Data":"1cf28177f9adbcf0a9c79a12471b1a52b0cd46040c717fee0d7d425a0e157447"} Dec 03 10:03:50 crc kubenswrapper[4573]: I1203 10:03:50.421020 4573 scope.go:117] "RemoveContainer" containerID="1cf28177f9adbcf0a9c79a12471b1a52b0cd46040c717fee0d7d425a0e157447" Dec 03 10:03:50 crc kubenswrapper[4573]: I1203 10:03:50.539508 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vzks7_must-gather-658jg_ccf6b300-52ec-486b-abd5-2dfb38ab3c1f/gather/0.log" Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.321803 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vzks7/must-gather-658jg"] Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.322512 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-vzks7/must-gather-658jg" podUID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerName="copy" containerID="cri-o://ec9663268efc71d193fe7d5b5b65c8b5f602b2577f73153a7cc4c0469e87cf3b" gracePeriod=2 Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.340042 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vzks7/must-gather-658jg"] Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.523158 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vzks7_must-gather-658jg_ccf6b300-52ec-486b-abd5-2dfb38ab3c1f/copy/0.log" Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.523621 4573 generic.go:334] "Generic (PLEG): container finished" podID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerID="ec9663268efc71d193fe7d5b5b65c8b5f602b2577f73153a7cc4c0469e87cf3b" exitCode=143 Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.791067 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vzks7_must-gather-658jg_ccf6b300-52ec-486b-abd5-2dfb38ab3c1f/copy/0.log" Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.791809 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.813002 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-must-gather-output\") pod \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\" (UID: \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\") " Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.813204 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbxmv\" (UniqueName: \"kubernetes.io/projected/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-kube-api-access-nbxmv\") pod \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\" (UID: \"ccf6b300-52ec-486b-abd5-2dfb38ab3c1f\") " Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.820536 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-kube-api-access-nbxmv" (OuterVolumeSpecName: "kube-api-access-nbxmv") pod "ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" (UID: "ccf6b300-52ec-486b-abd5-2dfb38ab3c1f"). InnerVolumeSpecName "kube-api-access-nbxmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:03:59 crc kubenswrapper[4573]: I1203 10:03:59.915355 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbxmv\" (UniqueName: \"kubernetes.io/projected/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-kube-api-access-nbxmv\") on node \"crc\" DevicePath \"\"" Dec 03 10:04:00 crc kubenswrapper[4573]: I1203 10:04:00.020897 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" (UID: "ccf6b300-52ec-486b-abd5-2dfb38ab3c1f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:04:00 crc kubenswrapper[4573]: I1203 10:04:00.022391 4573 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 10:04:00 crc kubenswrapper[4573]: I1203 10:04:00.046385 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" path="/var/lib/kubelet/pods/ccf6b300-52ec-486b-abd5-2dfb38ab3c1f/volumes" Dec 03 10:04:00 crc kubenswrapper[4573]: I1203 10:04:00.533536 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vzks7_must-gather-658jg_ccf6b300-52ec-486b-abd5-2dfb38ab3c1f/copy/0.log" Dec 03 10:04:00 crc kubenswrapper[4573]: I1203 10:04:00.534182 4573 scope.go:117] "RemoveContainer" containerID="ec9663268efc71d193fe7d5b5b65c8b5f602b2577f73153a7cc4c0469e87cf3b" Dec 03 10:04:00 crc kubenswrapper[4573]: I1203 10:04:00.534261 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vzks7/must-gather-658jg" Dec 03 10:04:00 crc kubenswrapper[4573]: I1203 10:04:00.556706 4573 scope.go:117] "RemoveContainer" containerID="1cf28177f9adbcf0a9c79a12471b1a52b0cd46040c717fee0d7d425a0e157447" Dec 03 10:04:09 crc kubenswrapper[4573]: I1203 10:04:09.889504 4573 scope.go:117] "RemoveContainer" containerID="8fe94cd29db6d826f6adb25629c474d5e69d3a400b2b5aabbd61e6215f0d5221" Dec 03 10:04:56 crc kubenswrapper[4573]: I1203 10:04:56.943727 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:04:56 crc kubenswrapper[4573]: I1203 10:04:56.944336 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:05:26 crc kubenswrapper[4573]: I1203 10:05:26.942965 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:05:26 crc kubenswrapper[4573]: I1203 10:05:26.943759 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:05:56 crc kubenswrapper[4573]: I1203 10:05:56.943194 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:05:56 crc kubenswrapper[4573]: I1203 10:05:56.945617 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:05:56 crc kubenswrapper[4573]: I1203 10:05:56.945772 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 10:05:56 crc kubenswrapper[4573]: I1203 10:05:56.946707 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 10:05:56 crc kubenswrapper[4573]: I1203 10:05:56.946871 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" gracePeriod=600 Dec 03 10:05:57 crc kubenswrapper[4573]: E1203 10:05:57.533260 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:05:57 crc kubenswrapper[4573]: I1203 10:05:57.705069 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" exitCode=0 Dec 03 10:05:57 crc kubenswrapper[4573]: I1203 10:05:57.705119 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2"} Dec 03 10:05:57 crc kubenswrapper[4573]: I1203 10:05:57.705172 4573 scope.go:117] "RemoveContainer" containerID="ed8849625f09788d6b1f0caa2c569eee66c1bad7cd396148adfae49d1b223065" Dec 03 10:05:57 crc kubenswrapper[4573]: I1203 10:05:57.706060 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:05:57 crc kubenswrapper[4573]: E1203 10:05:57.706320 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:06:08 crc kubenswrapper[4573]: I1203 10:06:08.033968 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:06:08 crc kubenswrapper[4573]: E1203 10:06:08.034915 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:06:22 crc kubenswrapper[4573]: I1203 10:06:22.031483 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:06:22 crc kubenswrapper[4573]: E1203 10:06:22.032158 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:06:36 crc kubenswrapper[4573]: I1203 10:06:36.031543 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:06:36 crc kubenswrapper[4573]: E1203 10:06:36.032420 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:06:50 crc kubenswrapper[4573]: I1203 10:06:50.075913 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:06:50 crc kubenswrapper[4573]: E1203 10:06:50.077207 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.663523 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6nkm/must-gather-wfn97"] Dec 03 10:07:01 crc kubenswrapper[4573]: E1203 10:07:01.664544 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerName="copy" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.664566 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerName="copy" Dec 03 10:07:01 crc kubenswrapper[4573]: E1203 10:07:01.664600 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerName="gather" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.664613 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerName="gather" Dec 03 10:07:01 crc kubenswrapper[4573]: E1203 10:07:01.664653 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerName="extract-utilities" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.664668 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerName="extract-utilities" Dec 03 10:07:01 crc kubenswrapper[4573]: E1203 10:07:01.664701 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerName="extract-content" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.664710 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerName="extract-content" Dec 03 10:07:01 crc kubenswrapper[4573]: E1203 10:07:01.664734 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerName="registry-server" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.664742 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerName="registry-server" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.664981 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerName="gather" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.665010 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="fecf37f5-06b8-4922-8136-a1453c5ed505" containerName="registry-server" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.665029 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf6b300-52ec-486b-abd5-2dfb38ab3c1f" containerName="copy" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.675760 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.678600 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t6nkm"/"openshift-service-ca.crt" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.678794 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t6nkm"/"kube-root-ca.crt" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.688883 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t6nkm/must-gather-wfn97"] Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.816777 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2521d5a5-1671-42d0-bdd1-180b57a9287b-must-gather-output\") pod \"must-gather-wfn97\" (UID: \"2521d5a5-1671-42d0-bdd1-180b57a9287b\") " pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.816902 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ntmp\" (UniqueName: \"kubernetes.io/projected/2521d5a5-1671-42d0-bdd1-180b57a9287b-kube-api-access-7ntmp\") pod \"must-gather-wfn97\" (UID: \"2521d5a5-1671-42d0-bdd1-180b57a9287b\") " pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.918407 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ntmp\" (UniqueName: \"kubernetes.io/projected/2521d5a5-1671-42d0-bdd1-180b57a9287b-kube-api-access-7ntmp\") pod \"must-gather-wfn97\" (UID: \"2521d5a5-1671-42d0-bdd1-180b57a9287b\") " pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.918779 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2521d5a5-1671-42d0-bdd1-180b57a9287b-must-gather-output\") pod \"must-gather-wfn97\" (UID: \"2521d5a5-1671-42d0-bdd1-180b57a9287b\") " pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.919323 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2521d5a5-1671-42d0-bdd1-180b57a9287b-must-gather-output\") pod \"must-gather-wfn97\" (UID: \"2521d5a5-1671-42d0-bdd1-180b57a9287b\") " pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.935410 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t6nkm"/"kube-root-ca.crt" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.947464 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t6nkm"/"openshift-service-ca.crt" Dec 03 10:07:01 crc kubenswrapper[4573]: I1203 10:07:01.975904 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ntmp\" (UniqueName: \"kubernetes.io/projected/2521d5a5-1671-42d0-bdd1-180b57a9287b-kube-api-access-7ntmp\") pod \"must-gather-wfn97\" (UID: \"2521d5a5-1671-42d0-bdd1-180b57a9287b\") " pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:07:02 crc kubenswrapper[4573]: I1203 10:07:02.005467 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:07:02 crc kubenswrapper[4573]: I1203 10:07:02.030947 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:07:02 crc kubenswrapper[4573]: E1203 10:07:02.031334 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:07:02 crc kubenswrapper[4573]: I1203 10:07:02.612115 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t6nkm/must-gather-wfn97"] Dec 03 10:07:03 crc kubenswrapper[4573]: I1203 10:07:03.444334 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/must-gather-wfn97" event={"ID":"2521d5a5-1671-42d0-bdd1-180b57a9287b","Type":"ContainerStarted","Data":"5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700"} Dec 03 10:07:03 crc kubenswrapper[4573]: I1203 10:07:03.444745 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/must-gather-wfn97" event={"ID":"2521d5a5-1671-42d0-bdd1-180b57a9287b","Type":"ContainerStarted","Data":"0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18"} Dec 03 10:07:03 crc kubenswrapper[4573]: I1203 10:07:03.444756 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/must-gather-wfn97" event={"ID":"2521d5a5-1671-42d0-bdd1-180b57a9287b","Type":"ContainerStarted","Data":"df0140ea55ca123a334108adc990ae769443b29a59e5ee1150764d46b7413805"} Dec 03 10:07:03 crc kubenswrapper[4573]: I1203 10:07:03.466523 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t6nkm/must-gather-wfn97" podStartSLOduration=2.466490909 podStartE2EDuration="2.466490909s" podCreationTimestamp="2025-12-03 10:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 10:07:03.462395177 +0000 UTC m=+5344.030774436" watchObservedRunningTime="2025-12-03 10:07:03.466490909 +0000 UTC m=+5344.034870178" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.271663 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-4c4r8"] Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.273352 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.275617 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t6nkm"/"default-dockercfg-6v84p" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.327171 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4djd\" (UniqueName: \"kubernetes.io/projected/2e61d3a5-455f-42f5-bf37-9f0cc3595371-kube-api-access-p4djd\") pod \"crc-debug-4c4r8\" (UID: \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\") " pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.327266 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e61d3a5-455f-42f5-bf37-9f0cc3595371-host\") pod \"crc-debug-4c4r8\" (UID: \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\") " pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.429351 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4djd\" (UniqueName: \"kubernetes.io/projected/2e61d3a5-455f-42f5-bf37-9f0cc3595371-kube-api-access-p4djd\") pod \"crc-debug-4c4r8\" (UID: \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\") " pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.429695 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e61d3a5-455f-42f5-bf37-9f0cc3595371-host\") pod \"crc-debug-4c4r8\" (UID: \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\") " pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.429769 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e61d3a5-455f-42f5-bf37-9f0cc3595371-host\") pod \"crc-debug-4c4r8\" (UID: \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\") " pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.733848 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4djd\" (UniqueName: \"kubernetes.io/projected/2e61d3a5-455f-42f5-bf37-9f0cc3595371-kube-api-access-p4djd\") pod \"crc-debug-4c4r8\" (UID: \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\") " pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:07:07 crc kubenswrapper[4573]: I1203 10:07:07.891806 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:07:07 crc kubenswrapper[4573]: W1203 10:07:07.920659 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e61d3a5_455f_42f5_bf37_9f0cc3595371.slice/crio-70c0241cacf7c6c7a14c09d8f8b40c82203c1c8aec4a4b6dd2727fc2662d9928 WatchSource:0}: Error finding container 70c0241cacf7c6c7a14c09d8f8b40c82203c1c8aec4a4b6dd2727fc2662d9928: Status 404 returned error can't find the container with id 70c0241cacf7c6c7a14c09d8f8b40c82203c1c8aec4a4b6dd2727fc2662d9928 Dec 03 10:07:08 crc kubenswrapper[4573]: I1203 10:07:08.511895 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" event={"ID":"2e61d3a5-455f-42f5-bf37-9f0cc3595371","Type":"ContainerStarted","Data":"70c0241cacf7c6c7a14c09d8f8b40c82203c1c8aec4a4b6dd2727fc2662d9928"} Dec 03 10:07:09 crc kubenswrapper[4573]: I1203 10:07:09.524541 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" event={"ID":"2e61d3a5-455f-42f5-bf37-9f0cc3595371","Type":"ContainerStarted","Data":"197f8764ca39f3c23275e56634ea58d910eb26ad0c9007007a1fb21d49ac1573"} Dec 03 10:07:13 crc kubenswrapper[4573]: I1203 10:07:13.030621 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:07:13 crc kubenswrapper[4573]: E1203 10:07:13.032343 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:07:24 crc kubenswrapper[4573]: I1203 10:07:24.030348 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:07:24 crc kubenswrapper[4573]: E1203 10:07:24.031230 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:07:39 crc kubenswrapper[4573]: I1203 10:07:39.031321 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:07:39 crc kubenswrapper[4573]: E1203 10:07:39.032544 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:07:50 crc kubenswrapper[4573]: I1203 10:07:50.047282 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:07:50 crc kubenswrapper[4573]: E1203 10:07:50.048017 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:08:00 crc kubenswrapper[4573]: I1203 10:08:00.065393 4573 generic.go:334] "Generic (PLEG): container finished" podID="2e61d3a5-455f-42f5-bf37-9f0cc3595371" containerID="197f8764ca39f3c23275e56634ea58d910eb26ad0c9007007a1fb21d49ac1573" exitCode=0 Dec 03 10:08:00 crc kubenswrapper[4573]: I1203 10:08:00.065880 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" event={"ID":"2e61d3a5-455f-42f5-bf37-9f0cc3595371","Type":"ContainerDied","Data":"197f8764ca39f3c23275e56634ea58d910eb26ad0c9007007a1fb21d49ac1573"} Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.216588 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.229375 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4djd\" (UniqueName: \"kubernetes.io/projected/2e61d3a5-455f-42f5-bf37-9f0cc3595371-kube-api-access-p4djd\") pod \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\" (UID: \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\") " Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.229474 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e61d3a5-455f-42f5-bf37-9f0cc3595371-host\") pod \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\" (UID: \"2e61d3a5-455f-42f5-bf37-9f0cc3595371\") " Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.229871 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e61d3a5-455f-42f5-bf37-9f0cc3595371-host" (OuterVolumeSpecName: "host") pod "2e61d3a5-455f-42f5-bf37-9f0cc3595371" (UID: "2e61d3a5-455f-42f5-bf37-9f0cc3595371"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.266325 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-4c4r8"] Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.278850 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-4c4r8"] Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.323469 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e61d3a5-455f-42f5-bf37-9f0cc3595371-kube-api-access-p4djd" (OuterVolumeSpecName: "kube-api-access-p4djd") pod "2e61d3a5-455f-42f5-bf37-9f0cc3595371" (UID: "2e61d3a5-455f-42f5-bf37-9f0cc3595371"). InnerVolumeSpecName "kube-api-access-p4djd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.331990 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4djd\" (UniqueName: \"kubernetes.io/projected/2e61d3a5-455f-42f5-bf37-9f0cc3595371-kube-api-access-p4djd\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:01 crc kubenswrapper[4573]: I1203 10:08:01.332028 4573 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e61d3a5-455f-42f5-bf37-9f0cc3595371-host\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.057538 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e61d3a5-455f-42f5-bf37-9f0cc3595371" path="/var/lib/kubelet/pods/2e61d3a5-455f-42f5-bf37-9f0cc3595371/volumes" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.109601 4573 scope.go:117] "RemoveContainer" containerID="197f8764ca39f3c23275e56634ea58d910eb26ad0c9007007a1fb21d49ac1573" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.109725 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-4c4r8" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.474669 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-rndsz"] Dec 03 10:08:02 crc kubenswrapper[4573]: E1203 10:08:02.475267 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e61d3a5-455f-42f5-bf37-9f0cc3595371" containerName="container-00" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.475278 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e61d3a5-455f-42f5-bf37-9f0cc3595371" containerName="container-00" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.475470 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e61d3a5-455f-42f5-bf37-9f0cc3595371" containerName="container-00" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.476133 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.481581 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t6nkm"/"default-dockercfg-6v84p" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.565462 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/561a7f13-f868-4147-8d28-f674b08af8c9-host\") pod \"crc-debug-rndsz\" (UID: \"561a7f13-f868-4147-8d28-f674b08af8c9\") " pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.565672 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgdcs\" (UniqueName: \"kubernetes.io/projected/561a7f13-f868-4147-8d28-f674b08af8c9-kube-api-access-jgdcs\") pod \"crc-debug-rndsz\" (UID: \"561a7f13-f868-4147-8d28-f674b08af8c9\") " pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.667766 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgdcs\" (UniqueName: \"kubernetes.io/projected/561a7f13-f868-4147-8d28-f674b08af8c9-kube-api-access-jgdcs\") pod \"crc-debug-rndsz\" (UID: \"561a7f13-f868-4147-8d28-f674b08af8c9\") " pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.668013 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/561a7f13-f868-4147-8d28-f674b08af8c9-host\") pod \"crc-debug-rndsz\" (UID: \"561a7f13-f868-4147-8d28-f674b08af8c9\") " pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.668121 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/561a7f13-f868-4147-8d28-f674b08af8c9-host\") pod \"crc-debug-rndsz\" (UID: \"561a7f13-f868-4147-8d28-f674b08af8c9\") " pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.724546 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgdcs\" (UniqueName: \"kubernetes.io/projected/561a7f13-f868-4147-8d28-f674b08af8c9-kube-api-access-jgdcs\") pod \"crc-debug-rndsz\" (UID: \"561a7f13-f868-4147-8d28-f674b08af8c9\") " pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:02 crc kubenswrapper[4573]: I1203 10:08:02.797085 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:03 crc kubenswrapper[4573]: I1203 10:08:03.118639 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/crc-debug-rndsz" event={"ID":"561a7f13-f868-4147-8d28-f674b08af8c9","Type":"ContainerStarted","Data":"6b4aeeefe8aa1254f70136441981f6378e363dd1380b4ff7c49b8c888dc62bca"} Dec 03 10:08:04 crc kubenswrapper[4573]: I1203 10:08:04.130314 4573 generic.go:334] "Generic (PLEG): container finished" podID="561a7f13-f868-4147-8d28-f674b08af8c9" containerID="b418b1c5d229cffa0dcb78723a82c7821ed23e422d0cb8a5782d6c225a3e06a7" exitCode=0 Dec 03 10:08:04 crc kubenswrapper[4573]: I1203 10:08:04.130417 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/crc-debug-rndsz" event={"ID":"561a7f13-f868-4147-8d28-f674b08af8c9","Type":"ContainerDied","Data":"b418b1c5d229cffa0dcb78723a82c7821ed23e422d0cb8a5782d6c225a3e06a7"} Dec 03 10:08:05 crc kubenswrapper[4573]: I1203 10:08:05.264858 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:05 crc kubenswrapper[4573]: I1203 10:08:05.418654 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgdcs\" (UniqueName: \"kubernetes.io/projected/561a7f13-f868-4147-8d28-f674b08af8c9-kube-api-access-jgdcs\") pod \"561a7f13-f868-4147-8d28-f674b08af8c9\" (UID: \"561a7f13-f868-4147-8d28-f674b08af8c9\") " Dec 03 10:08:05 crc kubenswrapper[4573]: I1203 10:08:05.418742 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/561a7f13-f868-4147-8d28-f674b08af8c9-host\") pod \"561a7f13-f868-4147-8d28-f674b08af8c9\" (UID: \"561a7f13-f868-4147-8d28-f674b08af8c9\") " Dec 03 10:08:05 crc kubenswrapper[4573]: I1203 10:08:05.419534 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561a7f13-f868-4147-8d28-f674b08af8c9-host" (OuterVolumeSpecName: "host") pod "561a7f13-f868-4147-8d28-f674b08af8c9" (UID: "561a7f13-f868-4147-8d28-f674b08af8c9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:08:05 crc kubenswrapper[4573]: I1203 10:08:05.438389 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561a7f13-f868-4147-8d28-f674b08af8c9-kube-api-access-jgdcs" (OuterVolumeSpecName: "kube-api-access-jgdcs") pod "561a7f13-f868-4147-8d28-f674b08af8c9" (UID: "561a7f13-f868-4147-8d28-f674b08af8c9"). InnerVolumeSpecName "kube-api-access-jgdcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:08:05 crc kubenswrapper[4573]: I1203 10:08:05.520939 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgdcs\" (UniqueName: \"kubernetes.io/projected/561a7f13-f868-4147-8d28-f674b08af8c9-kube-api-access-jgdcs\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:05 crc kubenswrapper[4573]: I1203 10:08:05.520985 4573 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/561a7f13-f868-4147-8d28-f674b08af8c9-host\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:06 crc kubenswrapper[4573]: I1203 10:08:06.030302 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:08:06 crc kubenswrapper[4573]: E1203 10:08:06.031024 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:08:06 crc kubenswrapper[4573]: I1203 10:08:06.156357 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-rndsz" Dec 03 10:08:06 crc kubenswrapper[4573]: I1203 10:08:06.156280 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/crc-debug-rndsz" event={"ID":"561a7f13-f868-4147-8d28-f674b08af8c9","Type":"ContainerDied","Data":"6b4aeeefe8aa1254f70136441981f6378e363dd1380b4ff7c49b8c888dc62bca"} Dec 03 10:08:06 crc kubenswrapper[4573]: I1203 10:08:06.157598 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b4aeeefe8aa1254f70136441981f6378e363dd1380b4ff7c49b8c888dc62bca" Dec 03 10:08:06 crc kubenswrapper[4573]: I1203 10:08:06.708565 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-rndsz"] Dec 03 10:08:06 crc kubenswrapper[4573]: I1203 10:08:06.715502 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-rndsz"] Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.027743 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-n26tk"] Dec 03 10:08:08 crc kubenswrapper[4573]: E1203 10:08:08.028528 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561a7f13-f868-4147-8d28-f674b08af8c9" containerName="container-00" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.028547 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="561a7f13-f868-4147-8d28-f674b08af8c9" containerName="container-00" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.028751 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="561a7f13-f868-4147-8d28-f674b08af8c9" containerName="container-00" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.029555 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.032324 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t6nkm"/"default-dockercfg-6v84p" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.041864 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561a7f13-f868-4147-8d28-f674b08af8c9" path="/var/lib/kubelet/pods/561a7f13-f868-4147-8d28-f674b08af8c9/volumes" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.151895 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-host\") pod \"crc-debug-n26tk\" (UID: \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\") " pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.151975 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcl9c\" (UniqueName: \"kubernetes.io/projected/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-kube-api-access-wcl9c\") pod \"crc-debug-n26tk\" (UID: \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\") " pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.257085 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-host\") pod \"crc-debug-n26tk\" (UID: \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\") " pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.257165 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcl9c\" (UniqueName: \"kubernetes.io/projected/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-kube-api-access-wcl9c\") pod \"crc-debug-n26tk\" (UID: \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\") " pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.257488 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-host\") pod \"crc-debug-n26tk\" (UID: \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\") " pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.301705 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcl9c\" (UniqueName: \"kubernetes.io/projected/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-kube-api-access-wcl9c\") pod \"crc-debug-n26tk\" (UID: \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\") " pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:08 crc kubenswrapper[4573]: I1203 10:08:08.349631 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:09 crc kubenswrapper[4573]: I1203 10:08:09.209485 4573 generic.go:334] "Generic (PLEG): container finished" podID="8e4eb5bc-4b35-406e-98a9-435fd5d9bd58" containerID="f5ac964a361a626eda39beb48b9d243820e9536a2928ff9411974ed630ae5171" exitCode=0 Dec 03 10:08:09 crc kubenswrapper[4573]: I1203 10:08:09.209564 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/crc-debug-n26tk" event={"ID":"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58","Type":"ContainerDied","Data":"f5ac964a361a626eda39beb48b9d243820e9536a2928ff9411974ed630ae5171"} Dec 03 10:08:09 crc kubenswrapper[4573]: I1203 10:08:09.209838 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/crc-debug-n26tk" event={"ID":"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58","Type":"ContainerStarted","Data":"3145849cfe9061c2332152dcd280d241fd2cad650ab723e5ffb55e279fea21c9"} Dec 03 10:08:09 crc kubenswrapper[4573]: I1203 10:08:09.248173 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-n26tk"] Dec 03 10:08:09 crc kubenswrapper[4573]: I1203 10:08:09.256005 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6nkm/crc-debug-n26tk"] Dec 03 10:08:10 crc kubenswrapper[4573]: I1203 10:08:10.321878 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:10 crc kubenswrapper[4573]: I1203 10:08:10.506653 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-host\") pod \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\" (UID: \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\") " Dec 03 10:08:10 crc kubenswrapper[4573]: I1203 10:08:10.506801 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-host" (OuterVolumeSpecName: "host") pod "8e4eb5bc-4b35-406e-98a9-435fd5d9bd58" (UID: "8e4eb5bc-4b35-406e-98a9-435fd5d9bd58"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 10:08:10 crc kubenswrapper[4573]: I1203 10:08:10.507198 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcl9c\" (UniqueName: \"kubernetes.io/projected/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-kube-api-access-wcl9c\") pod \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\" (UID: \"8e4eb5bc-4b35-406e-98a9-435fd5d9bd58\") " Dec 03 10:08:10 crc kubenswrapper[4573]: I1203 10:08:10.507604 4573 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-host\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:10 crc kubenswrapper[4573]: I1203 10:08:10.527481 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-kube-api-access-wcl9c" (OuterVolumeSpecName: "kube-api-access-wcl9c") pod "8e4eb5bc-4b35-406e-98a9-435fd5d9bd58" (UID: "8e4eb5bc-4b35-406e-98a9-435fd5d9bd58"). InnerVolumeSpecName "kube-api-access-wcl9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:08:10 crc kubenswrapper[4573]: I1203 10:08:10.610186 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcl9c\" (UniqueName: \"kubernetes.io/projected/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58-kube-api-access-wcl9c\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:11 crc kubenswrapper[4573]: I1203 10:08:11.227775 4573 scope.go:117] "RemoveContainer" containerID="f5ac964a361a626eda39beb48b9d243820e9536a2928ff9411974ed630ae5171" Dec 03 10:08:11 crc kubenswrapper[4573]: I1203 10:08:11.227845 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/crc-debug-n26tk" Dec 03 10:08:12 crc kubenswrapper[4573]: I1203 10:08:12.044320 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e4eb5bc-4b35-406e-98a9-435fd5d9bd58" path="/var/lib/kubelet/pods/8e4eb5bc-4b35-406e-98a9-435fd5d9bd58/volumes" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.574730 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7gfcc"] Dec 03 10:08:17 crc kubenswrapper[4573]: E1203 10:08:17.576287 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4eb5bc-4b35-406e-98a9-435fd5d9bd58" containerName="container-00" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.576311 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4eb5bc-4b35-406e-98a9-435fd5d9bd58" containerName="container-00" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.576694 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e4eb5bc-4b35-406e-98a9-435fd5d9bd58" containerName="container-00" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.578988 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.598670 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7gfcc"] Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.599936 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ptsf\" (UniqueName: \"kubernetes.io/projected/e82d7475-e9dd-40db-9238-3fab90a01dd9-kube-api-access-4ptsf\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.599983 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-catalog-content\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.600099 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-utilities\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.703205 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-utilities\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.703344 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ptsf\" (UniqueName: \"kubernetes.io/projected/e82d7475-e9dd-40db-9238-3fab90a01dd9-kube-api-access-4ptsf\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.703389 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-catalog-content\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.704035 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-catalog-content\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.704468 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-utilities\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.735123 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ptsf\" (UniqueName: \"kubernetes.io/projected/e82d7475-e9dd-40db-9238-3fab90a01dd9-kube-api-access-4ptsf\") pod \"certified-operators-7gfcc\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:17 crc kubenswrapper[4573]: I1203 10:08:17.900975 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:18 crc kubenswrapper[4573]: I1203 10:08:18.030524 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:08:18 crc kubenswrapper[4573]: E1203 10:08:18.030954 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:08:18 crc kubenswrapper[4573]: I1203 10:08:18.472693 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7gfcc"] Dec 03 10:08:19 crc kubenswrapper[4573]: I1203 10:08:19.306513 4573 generic.go:334] "Generic (PLEG): container finished" podID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerID="87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f" exitCode=0 Dec 03 10:08:19 crc kubenswrapper[4573]: I1203 10:08:19.306625 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gfcc" event={"ID":"e82d7475-e9dd-40db-9238-3fab90a01dd9","Type":"ContainerDied","Data":"87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f"} Dec 03 10:08:19 crc kubenswrapper[4573]: I1203 10:08:19.306844 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gfcc" event={"ID":"e82d7475-e9dd-40db-9238-3fab90a01dd9","Type":"ContainerStarted","Data":"b740d66ea10165447526be7cf92db0a0cfd3d138860256707e5c365e2f2e46bd"} Dec 03 10:08:19 crc kubenswrapper[4573]: I1203 10:08:19.309007 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 10:08:21 crc kubenswrapper[4573]: I1203 10:08:21.326493 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gfcc" event={"ID":"e82d7475-e9dd-40db-9238-3fab90a01dd9","Type":"ContainerStarted","Data":"5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569"} Dec 03 10:08:22 crc kubenswrapper[4573]: I1203 10:08:22.338521 4573 generic.go:334] "Generic (PLEG): container finished" podID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerID="5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569" exitCode=0 Dec 03 10:08:22 crc kubenswrapper[4573]: I1203 10:08:22.338780 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gfcc" event={"ID":"e82d7475-e9dd-40db-9238-3fab90a01dd9","Type":"ContainerDied","Data":"5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569"} Dec 03 10:08:23 crc kubenswrapper[4573]: I1203 10:08:23.349426 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gfcc" event={"ID":"e82d7475-e9dd-40db-9238-3fab90a01dd9","Type":"ContainerStarted","Data":"556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d"} Dec 03 10:08:23 crc kubenswrapper[4573]: I1203 10:08:23.375626 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7gfcc" podStartSLOduration=2.791212559 podStartE2EDuration="6.375604018s" podCreationTimestamp="2025-12-03 10:08:17 +0000 UTC" firstStartedPulling="2025-12-03 10:08:19.308730473 +0000 UTC m=+5419.877109732" lastFinishedPulling="2025-12-03 10:08:22.893121932 +0000 UTC m=+5423.461501191" observedRunningTime="2025-12-03 10:08:23.366827909 +0000 UTC m=+5423.935207178" watchObservedRunningTime="2025-12-03 10:08:23.375604018 +0000 UTC m=+5423.943983277" Dec 03 10:08:27 crc kubenswrapper[4573]: I1203 10:08:27.901809 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:27 crc kubenswrapper[4573]: I1203 10:08:27.902069 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:27 crc kubenswrapper[4573]: I1203 10:08:27.956569 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:28 crc kubenswrapper[4573]: I1203 10:08:28.499312 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:28 crc kubenswrapper[4573]: I1203 10:08:28.557473 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7gfcc"] Dec 03 10:08:30 crc kubenswrapper[4573]: I1203 10:08:30.439494 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7gfcc" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerName="registry-server" containerID="cri-o://556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d" gracePeriod=2 Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.031470 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:08:31 crc kubenswrapper[4573]: E1203 10:08:31.032477 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.431442 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.455363 4573 generic.go:334] "Generic (PLEG): container finished" podID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerID="556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d" exitCode=0 Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.455418 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gfcc" event={"ID":"e82d7475-e9dd-40db-9238-3fab90a01dd9","Type":"ContainerDied","Data":"556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d"} Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.455452 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7gfcc" event={"ID":"e82d7475-e9dd-40db-9238-3fab90a01dd9","Type":"ContainerDied","Data":"b740d66ea10165447526be7cf92db0a0cfd3d138860256707e5c365e2f2e46bd"} Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.455475 4573 scope.go:117] "RemoveContainer" containerID="556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.455641 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7gfcc" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.493281 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-catalog-content\") pod \"e82d7475-e9dd-40db-9238-3fab90a01dd9\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.493339 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-utilities\") pod \"e82d7475-e9dd-40db-9238-3fab90a01dd9\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.493572 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ptsf\" (UniqueName: \"kubernetes.io/projected/e82d7475-e9dd-40db-9238-3fab90a01dd9-kube-api-access-4ptsf\") pod \"e82d7475-e9dd-40db-9238-3fab90a01dd9\" (UID: \"e82d7475-e9dd-40db-9238-3fab90a01dd9\") " Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.496459 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-utilities" (OuterVolumeSpecName: "utilities") pod "e82d7475-e9dd-40db-9238-3fab90a01dd9" (UID: "e82d7475-e9dd-40db-9238-3fab90a01dd9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.503511 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e82d7475-e9dd-40db-9238-3fab90a01dd9-kube-api-access-4ptsf" (OuterVolumeSpecName: "kube-api-access-4ptsf") pod "e82d7475-e9dd-40db-9238-3fab90a01dd9" (UID: "e82d7475-e9dd-40db-9238-3fab90a01dd9"). InnerVolumeSpecName "kube-api-access-4ptsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.520568 4573 scope.go:117] "RemoveContainer" containerID="5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.549757 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e82d7475-e9dd-40db-9238-3fab90a01dd9" (UID: "e82d7475-e9dd-40db-9238-3fab90a01dd9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.562364 4573 scope.go:117] "RemoveContainer" containerID="87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.596436 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.596476 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e82d7475-e9dd-40db-9238-3fab90a01dd9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.596491 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ptsf\" (UniqueName: \"kubernetes.io/projected/e82d7475-e9dd-40db-9238-3fab90a01dd9-kube-api-access-4ptsf\") on node \"crc\" DevicePath \"\"" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.607803 4573 scope.go:117] "RemoveContainer" containerID="556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d" Dec 03 10:08:31 crc kubenswrapper[4573]: E1203 10:08:31.608375 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d\": container with ID starting with 556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d not found: ID does not exist" containerID="556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.608416 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d"} err="failed to get container status \"556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d\": rpc error: code = NotFound desc = could not find container \"556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d\": container with ID starting with 556aa32c9986576fec31aef63894312253c54234502b5010d642ccfe7360913d not found: ID does not exist" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.608444 4573 scope.go:117] "RemoveContainer" containerID="5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569" Dec 03 10:08:31 crc kubenswrapper[4573]: E1203 10:08:31.608853 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569\": container with ID starting with 5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569 not found: ID does not exist" containerID="5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.608883 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569"} err="failed to get container status \"5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569\": rpc error: code = NotFound desc = could not find container \"5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569\": container with ID starting with 5408c2a60b0277ae4a66955cadb29d756eef24b09b0eddaf0ac422beea69f569 not found: ID does not exist" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.608905 4573 scope.go:117] "RemoveContainer" containerID="87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f" Dec 03 10:08:31 crc kubenswrapper[4573]: E1203 10:08:31.609293 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f\": container with ID starting with 87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f not found: ID does not exist" containerID="87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.609321 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f"} err="failed to get container status \"87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f\": rpc error: code = NotFound desc = could not find container \"87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f\": container with ID starting with 87be65f410e2d4698ff3f5e927e740ffb61671496c3901d48c01086a51c3283f not found: ID does not exist" Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.799120 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7gfcc"] Dec 03 10:08:31 crc kubenswrapper[4573]: I1203 10:08:31.809310 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7gfcc"] Dec 03 10:08:32 crc kubenswrapper[4573]: I1203 10:08:32.044222 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" path="/var/lib/kubelet/pods/e82d7475-e9dd-40db-9238-3fab90a01dd9/volumes" Dec 03 10:08:42 crc kubenswrapper[4573]: I1203 10:08:42.030884 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:08:42 crc kubenswrapper[4573]: E1203 10:08:42.031486 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:08:52 crc kubenswrapper[4573]: I1203 10:08:52.611038 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-847b4dbb6-dsqmk_6907cde0-9a5e-432e-bc68-55b357a514d9/barbican-api/0.log" Dec 03 10:08:52 crc kubenswrapper[4573]: I1203 10:08:52.908793 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-847b4dbb6-dsqmk_6907cde0-9a5e-432e-bc68-55b357a514d9/barbican-api-log/0.log" Dec 03 10:08:52 crc kubenswrapper[4573]: I1203 10:08:52.910687 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c965d87c4-dmtfc_e954edc8-66c8-4f8f-a255-b8c5d810aa1a/barbican-keystone-listener-log/0.log" Dec 03 10:08:52 crc kubenswrapper[4573]: I1203 10:08:52.933918 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5c965d87c4-dmtfc_e954edc8-66c8-4f8f-a255-b8c5d810aa1a/barbican-keystone-listener/0.log" Dec 03 10:08:53 crc kubenswrapper[4573]: I1203 10:08:53.198417 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-cd9d86d7-kn6nh_865e87dd-1725-4932-9566-dbfbf7b85a60/barbican-worker/0.log" Dec 03 10:08:53 crc kubenswrapper[4573]: I1203 10:08:53.221830 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-cd9d86d7-kn6nh_865e87dd-1725-4932-9566-dbfbf7b85a60/barbican-worker-log/0.log" Dec 03 10:08:53 crc kubenswrapper[4573]: I1203 10:08:53.462569 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-55d4r_3dc88703-275c-419a-961e-9c034464b6cb/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:08:53 crc kubenswrapper[4573]: I1203 10:08:53.501074 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3601284e-acb1-4e60-a4c1-18c404dc4072/ceilometer-notification-agent/0.log" Dec 03 10:08:53 crc kubenswrapper[4573]: I1203 10:08:53.590310 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3601284e-acb1-4e60-a4c1-18c404dc4072/ceilometer-central-agent/0.log" Dec 03 10:08:53 crc kubenswrapper[4573]: I1203 10:08:53.741869 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3601284e-acb1-4e60-a4c1-18c404dc4072/proxy-httpd/0.log" Dec 03 10:08:53 crc kubenswrapper[4573]: I1203 10:08:53.774425 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3601284e-acb1-4e60-a4c1-18c404dc4072/sg-core/0.log" Dec 03 10:08:53 crc kubenswrapper[4573]: I1203 10:08:53.965917 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3f75b291-5f0e-422f-a640-a3a231c400ee/cinder-api/0.log" Dec 03 10:08:54 crc kubenswrapper[4573]: I1203 10:08:54.352488 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3f75b291-5f0e-422f-a640-a3a231c400ee/cinder-api-log/0.log" Dec 03 10:08:54 crc kubenswrapper[4573]: I1203 10:08:54.419845 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_20097614-b580-49bf-a8c7-04f371825b9a/cinder-scheduler/0.log" Dec 03 10:08:54 crc kubenswrapper[4573]: I1203 10:08:54.605469 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_20097614-b580-49bf-a8c7-04f371825b9a/probe/0.log" Dec 03 10:08:54 crc kubenswrapper[4573]: I1203 10:08:54.641086 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-qxzv8_ab4c1175-21f3-43e4-a7b1-3d957b6d6fd5/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:08:54 crc kubenswrapper[4573]: I1203 10:08:54.808647 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-swdkm_6ea550b6-a03c-42ff-b766-073b5ca7bf15/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:08:54 crc kubenswrapper[4573]: I1203 10:08:54.841835 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d47554775-5hw5x_eb528c46-bf42-4870-850c-dcf96e5a59b9/init/0.log" Dec 03 10:08:55 crc kubenswrapper[4573]: I1203 10:08:55.186986 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d47554775-5hw5x_eb528c46-bf42-4870-850c-dcf96e5a59b9/init/0.log" Dec 03 10:08:55 crc kubenswrapper[4573]: I1203 10:08:55.322734 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-6kp67_960face3-051e-4d24-b7ae-8e8b0f13d14e/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:08:55 crc kubenswrapper[4573]: I1203 10:08:55.423431 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6d47554775-5hw5x_eb528c46-bf42-4870-850c-dcf96e5a59b9/dnsmasq-dns/0.log" Dec 03 10:08:55 crc kubenswrapper[4573]: I1203 10:08:55.586436 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_36bf5497-62bd-4e75-99e8-0a0272cd644a/glance-httpd/0.log" Dec 03 10:08:55 crc kubenswrapper[4573]: I1203 10:08:55.595620 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_36bf5497-62bd-4e75-99e8-0a0272cd644a/glance-log/0.log" Dec 03 10:08:55 crc kubenswrapper[4573]: I1203 10:08:55.785884 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2/glance-log/0.log" Dec 03 10:08:55 crc kubenswrapper[4573]: I1203 10:08:55.808134 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_10c2c2c5-a6d3-4ace-b5c2-4eadd26f21a2/glance-httpd/0.log" Dec 03 10:08:56 crc kubenswrapper[4573]: I1203 10:08:56.029965 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:08:56 crc kubenswrapper[4573]: E1203 10:08:56.030204 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:08:56 crc kubenswrapper[4573]: I1203 10:08:56.050320 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-66f99c579b-gh7sh_b78e2453-adc3-4139-97c5-9f8a977df1ee/horizon/2.log" Dec 03 10:08:56 crc kubenswrapper[4573]: I1203 10:08:56.177784 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-66f99c579b-gh7sh_b78e2453-adc3-4139-97c5-9f8a977df1ee/horizon/1.log" Dec 03 10:08:56 crc kubenswrapper[4573]: I1203 10:08:56.688153 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-tsz8p_66bd3982-db80-45dc-a985-5b2f05a0e257/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:08:57 crc kubenswrapper[4573]: I1203 10:08:57.115978 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412601-xpbj9_928d502e-8067-4583-a118-206953580db2/keystone-cron/0.log" Dec 03 10:08:57 crc kubenswrapper[4573]: I1203 10:08:57.139200 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-cg5z8_2e6f10c4-0f5e-46d5-82e6-45e3a58cc79b/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:08:57 crc kubenswrapper[4573]: I1203 10:08:57.378979 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_e50ae89c-c3c5-44a5-9222-b14ddbadc294/kube-state-metrics/0.log" Dec 03 10:08:57 crc kubenswrapper[4573]: I1203 10:08:57.628899 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-xbc4d_022b0b02-d8bb-4560-af50-5d048f4ec520/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:08:58 crc kubenswrapper[4573]: I1203 10:08:58.193983 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-66f99c579b-gh7sh_b78e2453-adc3-4139-97c5-9f8a977df1ee/horizon-log/0.log" Dec 03 10:08:58 crc kubenswrapper[4573]: I1203 10:08:58.639751 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cfccd4555-2shp6_4434273d-ce2e-4ecd-820d-7f2ecf35bca7/neutron-httpd/0.log" Dec 03 10:08:58 crc kubenswrapper[4573]: I1203 10:08:58.734482 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xjg44_281980d3-9cd5-4f71-b6a1-8602818bc836/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:08:58 crc kubenswrapper[4573]: I1203 10:08:58.803588 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cfccd4555-2shp6_4434273d-ce2e-4ecd-820d-7f2ecf35bca7/neutron-api/0.log" Dec 03 10:08:58 crc kubenswrapper[4573]: I1203 10:08:58.853401 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-f56bd7d55-5qgxb_06a92024-3b18-43c5-a4aa-22cec1d43ff3/keystone-api/0.log" Dec 03 10:08:59 crc kubenswrapper[4573]: I1203 10:08:59.766799 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a96a7087-9cec-427a-b088-0d900045bde3/nova-cell0-conductor-conductor/0.log" Dec 03 10:09:00 crc kubenswrapper[4573]: I1203 10:09:00.115736 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2227cd84-de1e-4b88-8712-dbd6471e6c67/nova-cell1-conductor-conductor/0.log" Dec 03 10:09:00 crc kubenswrapper[4573]: I1203 10:09:00.415438 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_a7d45736-034b-4537-ba45-dd60f3bc3e7a/memcached/0.log" Dec 03 10:09:00 crc kubenswrapper[4573]: I1203 10:09:00.478977 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_192989bd-f8d0-43d2-9267-3022e451041c/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 10:09:00 crc kubenswrapper[4573]: I1203 10:09:00.528875 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b06177f1-1c12-41d9-b7f8-7df5e6986872/nova-api-log/0.log" Dec 03 10:09:00 crc kubenswrapper[4573]: I1203 10:09:00.731648 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-jc6rt_5847f67b-3130-4fa9-831f-bcfe0c27ad48/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:09:00 crc kubenswrapper[4573]: I1203 10:09:00.905224 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f233cf9b-b14d-4eee-8f84-3fcb5d85f571/nova-metadata-log/0.log" Dec 03 10:09:01 crc kubenswrapper[4573]: I1203 10:09:01.024514 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b06177f1-1c12-41d9-b7f8-7df5e6986872/nova-api-api/0.log" Dec 03 10:09:01 crc kubenswrapper[4573]: I1203 10:09:01.289030 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d/mysql-bootstrap/0.log" Dec 03 10:09:01 crc kubenswrapper[4573]: I1203 10:09:01.536012 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_56f569d8-50e2-4177-bec8-8225de477357/nova-scheduler-scheduler/0.log" Dec 03 10:09:01 crc kubenswrapper[4573]: I1203 10:09:01.733574 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d/galera/0.log" Dec 03 10:09:01 crc kubenswrapper[4573]: I1203 10:09:01.798000 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9671b52a-6c34-4d7d-ab90-fa5b3ef4f89d/mysql-bootstrap/0.log" Dec 03 10:09:01 crc kubenswrapper[4573]: I1203 10:09:01.931150 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96/mysql-bootstrap/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.055722 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96/mysql-bootstrap/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.129659 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_6c6e55a3-e2b9-4cfc-afa0-af3b9903ec96/galera/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.272722 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_4357c5a1-7325-4bff-8a79-4ffd70879b1c/openstackclient/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.466891 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bj74b_33a7f303-d65a-4546-bbb5-1e223d48d847/ovn-controller/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.475485 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f233cf9b-b14d-4eee-8f84-3fcb5d85f571/nova-metadata-metadata/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.655667 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-rrcpn_6a19cac1-e9f7-4f7d-b356-3263c1b8c446/openstack-network-exporter/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.717007 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jbx79_651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8/ovsdb-server-init/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.908189 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jbx79_651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8/ovsdb-server-init/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.968626 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jbx79_651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8/ovs-vswitchd/0.log" Dec 03 10:09:02 crc kubenswrapper[4573]: I1203 10:09:02.996790 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-jbx79_651ac0e8-f9bf-4b2c-bbd5-b642b6ca49c8/ovsdb-server/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.021947 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-vq5qw_633c66a3-1e25-4f81-9eee-2cb09bf54855/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.198364 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1999a198-5df7-40fc-ba0f-5b650b3027a6/openstack-network-exporter/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.247783 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1999a198-5df7-40fc-ba0f-5b650b3027a6/ovn-northd/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.263869 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1604082b-b77f-4df2-bea3-ab7803b3e5ec/openstack-network-exporter/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.389435 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1604082b-b77f-4df2-bea3-ab7803b3e5ec/ovsdbserver-nb/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.451404 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f5af474c-26c9-4877-989a-911cd77862ab/openstack-network-exporter/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.494823 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f5af474c-26c9-4877-989a-911cd77862ab/ovsdbserver-sb/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.828429 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-85bbcf664-5fm4n_b292e81d-5ded-4fff-bb9e-c764c395609b/placement-api/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.938076 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f038afb-2485-4e6e-8e0c-99ec178f4268/setup-container/0.log" Dec 03 10:09:03 crc kubenswrapper[4573]: I1203 10:09:03.989037 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-85bbcf664-5fm4n_b292e81d-5ded-4fff-bb9e-c764c395609b/placement-log/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.099804 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f038afb-2485-4e6e-8e0c-99ec178f4268/setup-container/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.122908 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5f038afb-2485-4e6e-8e0c-99ec178f4268/rabbitmq/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.244686 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fb1808-0ac8-4096-abff-76c44b1245f8/setup-container/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.384382 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-c5tzp_fb9f04b9-8dd3-4776-8dec-c469a94759c2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.498121 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fb1808-0ac8-4096-abff-76c44b1245f8/setup-container/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.518086 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_70fb1808-0ac8-4096-abff-76c44b1245f8/rabbitmq/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.646027 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-lbctc_609d7944-6cf6-47c4-a1f2-2437c57703d2/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.751673 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-glhs7_dd487e4c-80de-48b3-aaeb-1a3c9d5eda74/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:09:04 crc kubenswrapper[4573]: I1203 10:09:04.757443 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-qnv8j_fdbae972-3f32-4c7b-8ac3-e2e7eb5dd619/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.040911 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-gc8nt_9c8d5c32-d73f-4f7f-b1ad-02ebdf28f0c3/ssh-known-hosts-edpm-deployment/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.204612 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-66d678ddfc-gvbk5_02bc648a-d9c8-4fdf-bf48-bc57a6b05386/proxy-server/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.218066 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-66d678ddfc-gvbk5_02bc648a-d9c8-4fdf-bf48-bc57a6b05386/proxy-httpd/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.546204 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-lzbxl_5ceed462-6691-4229-b764-b25eb42bbf86/swift-ring-rebalance/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.625197 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/account-auditor/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.673572 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/account-replicator/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.683785 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/account-reaper/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.803927 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/account-server/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.863662 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/container-auditor/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.887159 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/container-replicator/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.915348 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/container-server/0.log" Dec 03 10:09:05 crc kubenswrapper[4573]: I1203 10:09:05.986432 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/container-updater/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.066446 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-expirer/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.068099 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-auditor/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.135434 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-replicator/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.175386 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-server/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.267017 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/object-updater/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.319800 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/swift-recon-cron/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.321511 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_053ecb51-80b3-4888-9384-8438ba293ae8/rsync/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.488559 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-tf5cc_fa5259c8-c545-4b45-a9c6-e0bca4d3aba9/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.595266 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_9c0dc4cc-36bd-40e6-99f0-4acb98a2ab8b/tempest-tests-tempest-tests-runner/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.635871 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f5b253a6-9fcb-432b-81d9-6a6b980367c7/test-operator-logs-container/0.log" Dec 03 10:09:06 crc kubenswrapper[4573]: I1203 10:09:06.818411 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4xfhf_8ebf9479-f67b-443c-8ae2-3bec8e719750/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 10:09:09 crc kubenswrapper[4573]: I1203 10:09:09.031225 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:09:09 crc kubenswrapper[4573]: E1203 10:09:09.031816 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:09:21 crc kubenswrapper[4573]: I1203 10:09:21.032674 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:09:21 crc kubenswrapper[4573]: E1203 10:09:21.033545 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:09:35 crc kubenswrapper[4573]: I1203 10:09:35.030878 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:09:35 crc kubenswrapper[4573]: E1203 10:09:35.032329 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:09:36 crc kubenswrapper[4573]: I1203 10:09:36.479504 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/util/0.log" Dec 03 10:09:36 crc kubenswrapper[4573]: I1203 10:09:36.694058 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/util/0.log" Dec 03 10:09:36 crc kubenswrapper[4573]: I1203 10:09:36.708833 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/pull/0.log" Dec 03 10:09:36 crc kubenswrapper[4573]: I1203 10:09:36.718691 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/pull/0.log" Dec 03 10:09:36 crc kubenswrapper[4573]: I1203 10:09:36.993432 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/pull/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.041981 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/extract/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.046356 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_646f2526e5c0154977d9ccbf52d7cac72400c0cdc63df62db2c6094ec1vgqhp_c09528b8-a5a1-490d-86ef-57027d97f282/util/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.186921 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-nx7mq_e27aea95-097f-4aa8-ae2b-7d212b8640b4/kube-rbac-proxy/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.336660 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-nx7mq_e27aea95-097f-4aa8-ae2b-7d212b8640b4/manager/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.437320 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-hb6l4_e0ae572b-e68c-48a0-8649-7eea884e61b0/kube-rbac-proxy/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.601960 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-hb6l4_e0ae572b-e68c-48a0-8649-7eea884e61b0/manager/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.683527 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-vl9xb_ee80479f-97a6-43ef-a026-d5bf1931d962/kube-rbac-proxy/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.704684 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-vl9xb_ee80479f-97a6-43ef-a026-d5bf1931d962/manager/0.log" Dec 03 10:09:37 crc kubenswrapper[4573]: I1203 10:09:37.921882 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-s8679_a855d891-6a41-477d-98d7-fcbba59aee28/kube-rbac-proxy/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.042953 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-s8679_a855d891-6a41-477d-98d7-fcbba59aee28/manager/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.166441 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gnlgf_ff1aaf12-2fcd-41d0-a575-a0e5447fc11f/kube-rbac-proxy/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.206032 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gnlgf_ff1aaf12-2fcd-41d0-a575-a0e5447fc11f/manager/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.321466 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2qnlp_622de9fb-c4ba-4727-8c22-5fcb36f39751/kube-rbac-proxy/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.519893 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2qnlp_622de9fb-c4ba-4727-8c22-5fcb36f39751/manager/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.532955 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gljjt_22bca7e1-22a7-4ee7-852b-25fddeb8fce3/kube-rbac-proxy/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.742318 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-gljjt_22bca7e1-22a7-4ee7-852b-25fddeb8fce3/manager/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.826017 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-wcbtn_9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b/kube-rbac-proxy/0.log" Dec 03 10:09:38 crc kubenswrapper[4573]: I1203 10:09:38.897535 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-wcbtn_9a4b6d1e-9887-45e0-9c8f-8ce38a92c28b/manager/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.001441 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-hdhrp_5aba2f13-92bd-4ea3-bfcb-7646909db04b/kube-rbac-proxy/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.198620 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-hdhrp_5aba2f13-92bd-4ea3-bfcb-7646909db04b/manager/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.306084 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qphvx_4613c040-3ca1-458c-8ff8-2b8858a7ad35/kube-rbac-proxy/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.376466 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-qphvx_4613c040-3ca1-458c-8ff8-2b8858a7ad35/manager/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.510226 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hl72w_a2888fa7-1f19-48c8-853e-a3952cc93ec6/kube-rbac-proxy/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.608300 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hl72w_a2888fa7-1f19-48c8-853e-a3952cc93ec6/manager/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.713580 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-r6qvb_9d6a307e-5cf2-4a23-921c-c5b562494cb0/kube-rbac-proxy/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.804693 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-r6qvb_9d6a307e-5cf2-4a23-921c-c5b562494cb0/manager/0.log" Dec 03 10:09:39 crc kubenswrapper[4573]: I1203 10:09:39.914847 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-sdrbz_fd426f33-5f03-4fec-bde4-e72e9c762762/kube-rbac-proxy/0.log" Dec 03 10:09:40 crc kubenswrapper[4573]: I1203 10:09:40.040977 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-sdrbz_fd426f33-5f03-4fec-bde4-e72e9c762762/manager/0.log" Dec 03 10:09:40 crc kubenswrapper[4573]: I1203 10:09:40.218745 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-bnztb_98fcc67c-249b-4700-8434-c8dd4e1646ae/manager/0.log" Dec 03 10:09:40 crc kubenswrapper[4573]: I1203 10:09:40.241926 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-bnztb_98fcc67c-249b-4700-8434-c8dd4e1646ae/kube-rbac-proxy/0.log" Dec 03 10:09:40 crc kubenswrapper[4573]: I1203 10:09:40.484200 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg_a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629/kube-rbac-proxy/0.log" Dec 03 10:09:40 crc kubenswrapper[4573]: I1203 10:09:40.521577 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd47ftvg_a0fafd12-f0eb-49a0-9d3a-7eefc2b5a629/manager/0.log" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.019434 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-f6f67fdf8-vgsnr_42fe335f-c3e5-4792-a174-2f5aeaa7d360/operator/0.log" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.180036 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-db2tp"] Dec 03 10:09:41 crc kubenswrapper[4573]: E1203 10:09:41.189651 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerName="extract-utilities" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.189691 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerName="extract-utilities" Dec 03 10:09:41 crc kubenswrapper[4573]: E1203 10:09:41.189756 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerName="extract-content" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.189767 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerName="extract-content" Dec 03 10:09:41 crc kubenswrapper[4573]: E1203 10:09:41.189781 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerName="registry-server" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.189788 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerName="registry-server" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.205029 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="e82d7475-e9dd-40db-9238-3fab90a01dd9" containerName="registry-server" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.211213 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-db2tp"] Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.211351 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.234719 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j6ps\" (UniqueName: \"kubernetes.io/projected/2b86feda-ecc4-45f7-934e-f1af251b73f8-kube-api-access-7j6ps\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.235025 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-catalog-content\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.235251 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-utilities\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.247183 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-h4tdp_1b63239f-176e-4a3b-ac7d-605be09bd269/registry-server/0.log" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.336554 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-catalog-content\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.336595 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j6ps\" (UniqueName: \"kubernetes.io/projected/2b86feda-ecc4-45f7-934e-f1af251b73f8-kube-api-access-7j6ps\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.336669 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-utilities\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.337295 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-utilities\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.337564 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-catalog-content\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.377473 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j6ps\" (UniqueName: \"kubernetes.io/projected/2b86feda-ecc4-45f7-934e-f1af251b73f8-kube-api-access-7j6ps\") pod \"redhat-marketplace-db2tp\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.583395 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.798908 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-d77h9_fd5e6711-117e-42b2-90e6-cfddf410315d/kube-rbac-proxy/0.log" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.914748 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-58f6d66c48-gwplc_5acea32c-fc74-40d8-b363-c64470ee880e/manager/0.log" Dec 03 10:09:41 crc kubenswrapper[4573]: I1203 10:09:41.925521 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-d77h9_fd5e6711-117e-42b2-90e6-cfddf410315d/manager/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.135820 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c5vc7_e94079e6-8701-48bf-ab58-1867fdf6e46e/kube-rbac-proxy/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.163556 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c5vc7_e94079e6-8701-48bf-ab58-1867fdf6e46e/manager/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.178912 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-db2tp"] Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.360750 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-jbwlq_b4436129-0e91-443b-8a97-61fa7f8cc2e4/operator/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.488932 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-nbfxj_57f767e9-7c3a-4707-9b76-ed715c297238/kube-rbac-proxy/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.504235 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-nbfxj_57f767e9-7c3a-4707-9b76-ed715c297238/manager/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.657016 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-jd6bv_8f65995e-e70a-48e4-9504-8aa85d43ea9e/kube-rbac-proxy/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.745785 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-jd6bv_8f65995e-e70a-48e4-9504-8aa85d43ea9e/manager/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.770850 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wdvlj_6535c22e-f476-4067-b1ab-7e0ed5c23360/kube-rbac-proxy/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.901789 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wdvlj_6535c22e-f476-4067-b1ab-7e0ed5c23360/manager/0.log" Dec 03 10:09:42 crc kubenswrapper[4573]: I1203 10:09:42.926031 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-l5vf4_e6f9febf-6db7-4ea8-8b96-7a939102be2a/kube-rbac-proxy/0.log" Dec 03 10:09:43 crc kubenswrapper[4573]: I1203 10:09:43.012125 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-l5vf4_e6f9febf-6db7-4ea8-8b96-7a939102be2a/manager/0.log" Dec 03 10:09:43 crc kubenswrapper[4573]: I1203 10:09:43.191573 4573 generic.go:334] "Generic (PLEG): container finished" podID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerID="fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d" exitCode=0 Dec 03 10:09:43 crc kubenswrapper[4573]: I1203 10:09:43.191634 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db2tp" event={"ID":"2b86feda-ecc4-45f7-934e-f1af251b73f8","Type":"ContainerDied","Data":"fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d"} Dec 03 10:09:43 crc kubenswrapper[4573]: I1203 10:09:43.191659 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db2tp" event={"ID":"2b86feda-ecc4-45f7-934e-f1af251b73f8","Type":"ContainerStarted","Data":"b494b9877555635175e7111f5ebfeb99b6a56de224ec5e40c917a7f8a9e56fd2"} Dec 03 10:09:44 crc kubenswrapper[4573]: I1203 10:09:44.204615 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db2tp" event={"ID":"2b86feda-ecc4-45f7-934e-f1af251b73f8","Type":"ContainerStarted","Data":"aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b"} Dec 03 10:09:45 crc kubenswrapper[4573]: I1203 10:09:45.215362 4573 generic.go:334] "Generic (PLEG): container finished" podID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerID="aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b" exitCode=0 Dec 03 10:09:45 crc kubenswrapper[4573]: I1203 10:09:45.215408 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db2tp" event={"ID":"2b86feda-ecc4-45f7-934e-f1af251b73f8","Type":"ContainerDied","Data":"aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b"} Dec 03 10:09:46 crc kubenswrapper[4573]: I1203 10:09:46.226435 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db2tp" event={"ID":"2b86feda-ecc4-45f7-934e-f1af251b73f8","Type":"ContainerStarted","Data":"899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39"} Dec 03 10:09:46 crc kubenswrapper[4573]: I1203 10:09:46.252568 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-db2tp" podStartSLOduration=2.794802856 podStartE2EDuration="5.25254874s" podCreationTimestamp="2025-12-03 10:09:41 +0000 UTC" firstStartedPulling="2025-12-03 10:09:43.193439033 +0000 UTC m=+5503.761818292" lastFinishedPulling="2025-12-03 10:09:45.651184917 +0000 UTC m=+5506.219564176" observedRunningTime="2025-12-03 10:09:46.241152429 +0000 UTC m=+5506.809531688" watchObservedRunningTime="2025-12-03 10:09:46.25254874 +0000 UTC m=+5506.820927999" Dec 03 10:09:50 crc kubenswrapper[4573]: I1203 10:09:50.037175 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:09:50 crc kubenswrapper[4573]: E1203 10:09:50.037934 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:09:51 crc kubenswrapper[4573]: I1203 10:09:51.583634 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:51 crc kubenswrapper[4573]: I1203 10:09:51.584443 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:51 crc kubenswrapper[4573]: I1203 10:09:51.639668 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:52 crc kubenswrapper[4573]: I1203 10:09:52.347352 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:52 crc kubenswrapper[4573]: I1203 10:09:52.400572 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-db2tp"] Dec 03 10:09:54 crc kubenswrapper[4573]: I1203 10:09:54.294611 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-db2tp" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerName="registry-server" containerID="cri-o://899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39" gracePeriod=2 Dec 03 10:09:54 crc kubenswrapper[4573]: I1203 10:09:54.864898 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.023486 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7j6ps\" (UniqueName: \"kubernetes.io/projected/2b86feda-ecc4-45f7-934e-f1af251b73f8-kube-api-access-7j6ps\") pod \"2b86feda-ecc4-45f7-934e-f1af251b73f8\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.023579 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-utilities\") pod \"2b86feda-ecc4-45f7-934e-f1af251b73f8\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.023703 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-catalog-content\") pod \"2b86feda-ecc4-45f7-934e-f1af251b73f8\" (UID: \"2b86feda-ecc4-45f7-934e-f1af251b73f8\") " Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.029770 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-utilities" (OuterVolumeSpecName: "utilities") pod "2b86feda-ecc4-45f7-934e-f1af251b73f8" (UID: "2b86feda-ecc4-45f7-934e-f1af251b73f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.031161 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b86feda-ecc4-45f7-934e-f1af251b73f8-kube-api-access-7j6ps" (OuterVolumeSpecName: "kube-api-access-7j6ps") pod "2b86feda-ecc4-45f7-934e-f1af251b73f8" (UID: "2b86feda-ecc4-45f7-934e-f1af251b73f8"). InnerVolumeSpecName "kube-api-access-7j6ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.041194 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b86feda-ecc4-45f7-934e-f1af251b73f8" (UID: "2b86feda-ecc4-45f7-934e-f1af251b73f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.126320 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7j6ps\" (UniqueName: \"kubernetes.io/projected/2b86feda-ecc4-45f7-934e-f1af251b73f8-kube-api-access-7j6ps\") on node \"crc\" DevicePath \"\"" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.126376 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.126391 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b86feda-ecc4-45f7-934e-f1af251b73f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.306215 4573 generic.go:334] "Generic (PLEG): container finished" podID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerID="899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39" exitCode=0 Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.306274 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db2tp" event={"ID":"2b86feda-ecc4-45f7-934e-f1af251b73f8","Type":"ContainerDied","Data":"899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39"} Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.306305 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-db2tp" event={"ID":"2b86feda-ecc4-45f7-934e-f1af251b73f8","Type":"ContainerDied","Data":"b494b9877555635175e7111f5ebfeb99b6a56de224ec5e40c917a7f8a9e56fd2"} Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.306328 4573 scope.go:117] "RemoveContainer" containerID="899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.306494 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-db2tp" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.346803 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-db2tp"] Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.348410 4573 scope.go:117] "RemoveContainer" containerID="aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.355828 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-db2tp"] Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.381823 4573 scope.go:117] "RemoveContainer" containerID="fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.428409 4573 scope.go:117] "RemoveContainer" containerID="899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39" Dec 03 10:09:55 crc kubenswrapper[4573]: E1203 10:09:55.428878 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39\": container with ID starting with 899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39 not found: ID does not exist" containerID="899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.428926 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39"} err="failed to get container status \"899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39\": rpc error: code = NotFound desc = could not find container \"899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39\": container with ID starting with 899778cf7c48c7b48bb1983d2ba23b6132ba766345d0b58d709e1294595c5c39 not found: ID does not exist" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.428955 4573 scope.go:117] "RemoveContainer" containerID="aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b" Dec 03 10:09:55 crc kubenswrapper[4573]: E1203 10:09:55.429234 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b\": container with ID starting with aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b not found: ID does not exist" containerID="aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.429272 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b"} err="failed to get container status \"aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b\": rpc error: code = NotFound desc = could not find container \"aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b\": container with ID starting with aeb95f0f105146661bc6db116a25be57a7d1cdc3d72d429af81627a68291240b not found: ID does not exist" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.429290 4573 scope.go:117] "RemoveContainer" containerID="fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d" Dec 03 10:09:55 crc kubenswrapper[4573]: E1203 10:09:55.429504 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d\": container with ID starting with fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d not found: ID does not exist" containerID="fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d" Dec 03 10:09:55 crc kubenswrapper[4573]: I1203 10:09:55.429525 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d"} err="failed to get container status \"fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d\": rpc error: code = NotFound desc = could not find container \"fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d\": container with ID starting with fe8c22d1bcb601a991f2b4e43626fb11bd4c15f8f57a6645e8bb628f2920aa4d not found: ID does not exist" Dec 03 10:09:56 crc kubenswrapper[4573]: I1203 10:09:56.043717 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" path="/var/lib/kubelet/pods/2b86feda-ecc4-45f7-934e-f1af251b73f8/volumes" Dec 03 10:10:03 crc kubenswrapper[4573]: I1203 10:10:03.031768 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:10:03 crc kubenswrapper[4573]: E1203 10:10:03.032828 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:10:04 crc kubenswrapper[4573]: I1203 10:10:04.753938 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-7mk94_192416c5-8025-4dd5-840b-9954ce7d009b/control-plane-machine-set-operator/0.log" Dec 03 10:10:04 crc kubenswrapper[4573]: I1203 10:10:04.912794 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8qfqs_284d9db2-04e5-4c39-b4cf-6f001de356c4/kube-rbac-proxy/0.log" Dec 03 10:10:04 crc kubenswrapper[4573]: I1203 10:10:04.930442 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8qfqs_284d9db2-04e5-4c39-b4cf-6f001de356c4/machine-api-operator/0.log" Dec 03 10:10:11 crc kubenswrapper[4573]: I1203 10:10:11.915032 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jklvl"] Dec 03 10:10:11 crc kubenswrapper[4573]: E1203 10:10:11.915909 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerName="registry-server" Dec 03 10:10:11 crc kubenswrapper[4573]: I1203 10:10:11.915922 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerName="registry-server" Dec 03 10:10:11 crc kubenswrapper[4573]: E1203 10:10:11.915938 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerName="extract-content" Dec 03 10:10:11 crc kubenswrapper[4573]: I1203 10:10:11.915944 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerName="extract-content" Dec 03 10:10:11 crc kubenswrapper[4573]: E1203 10:10:11.915964 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerName="extract-utilities" Dec 03 10:10:11 crc kubenswrapper[4573]: I1203 10:10:11.915970 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerName="extract-utilities" Dec 03 10:10:11 crc kubenswrapper[4573]: I1203 10:10:11.916256 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b86feda-ecc4-45f7-934e-f1af251b73f8" containerName="registry-server" Dec 03 10:10:11 crc kubenswrapper[4573]: I1203 10:10:11.919723 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:11 crc kubenswrapper[4573]: I1203 10:10:11.938356 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jklvl"] Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.084569 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-catalog-content\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.084662 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-utilities\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.084736 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h99cp\" (UniqueName: \"kubernetes.io/projected/16882f5a-571a-471e-92e0-7ff4307fa6c5-kube-api-access-h99cp\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.187069 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-catalog-content\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.187134 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-utilities\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.187196 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h99cp\" (UniqueName: \"kubernetes.io/projected/16882f5a-571a-471e-92e0-7ff4307fa6c5-kube-api-access-h99cp\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.188300 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-catalog-content\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.188565 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-utilities\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.210002 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h99cp\" (UniqueName: \"kubernetes.io/projected/16882f5a-571a-471e-92e0-7ff4307fa6c5-kube-api-access-h99cp\") pod \"redhat-operators-jklvl\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.254438 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:12 crc kubenswrapper[4573]: I1203 10:10:12.736971 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jklvl"] Dec 03 10:10:13 crc kubenswrapper[4573]: I1203 10:10:13.491451 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jklvl" event={"ID":"16882f5a-571a-471e-92e0-7ff4307fa6c5","Type":"ContainerStarted","Data":"7d12851de5eb1fb67110b03025c97df52eda495c9f42f246fb4eccac1600f964"} Dec 03 10:10:14 crc kubenswrapper[4573]: I1203 10:10:14.030322 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:10:14 crc kubenswrapper[4573]: E1203 10:10:14.030693 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:10:14 crc kubenswrapper[4573]: I1203 10:10:14.504330 4573 generic.go:334] "Generic (PLEG): container finished" podID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerID="1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991" exitCode=0 Dec 03 10:10:14 crc kubenswrapper[4573]: I1203 10:10:14.504377 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jklvl" event={"ID":"16882f5a-571a-471e-92e0-7ff4307fa6c5","Type":"ContainerDied","Data":"1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991"} Dec 03 10:10:16 crc kubenswrapper[4573]: I1203 10:10:16.540203 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jklvl" event={"ID":"16882f5a-571a-471e-92e0-7ff4307fa6c5","Type":"ContainerStarted","Data":"40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc"} Dec 03 10:10:19 crc kubenswrapper[4573]: I1203 10:10:19.394335 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-h9bp4_ee0f7af7-48fb-433a-8f6b-801fd53f331d/cert-manager-controller/0.log" Dec 03 10:10:19 crc kubenswrapper[4573]: I1203 10:10:19.698962 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-4226f_9d7c2bf5-9ddc-4b6e-9aa3-f5cef849aedf/cert-manager-cainjector/0.log" Dec 03 10:10:19 crc kubenswrapper[4573]: I1203 10:10:19.795716 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-twvqd_f4edafce-1ec1-4ada-9861-c36ce8c3906f/cert-manager-webhook/0.log" Dec 03 10:10:22 crc kubenswrapper[4573]: I1203 10:10:22.362263 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 10:10:22 crc kubenswrapper[4573]: I1203 10:10:22.362632 4573 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 10:10:22 crc kubenswrapper[4573]: I1203 10:10:22.362333 4573 patch_prober.go:28] interesting pod/router-default-5444994796-8tpmr container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 10:10:22 crc kubenswrapper[4573]: I1203 10:10:22.362742 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-8tpmr" podUID="f5a7efaa-3253-40d5-9e1e-59482378cd8c" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 10:10:23 crc kubenswrapper[4573]: I1203 10:10:23.633370 4573 generic.go:334] "Generic (PLEG): container finished" podID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerID="40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc" exitCode=0 Dec 03 10:10:23 crc kubenswrapper[4573]: I1203 10:10:23.633459 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jklvl" event={"ID":"16882f5a-571a-471e-92e0-7ff4307fa6c5","Type":"ContainerDied","Data":"40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc"} Dec 03 10:10:24 crc kubenswrapper[4573]: I1203 10:10:24.645436 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jklvl" event={"ID":"16882f5a-571a-471e-92e0-7ff4307fa6c5","Type":"ContainerStarted","Data":"8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455"} Dec 03 10:10:24 crc kubenswrapper[4573]: I1203 10:10:24.670792 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jklvl" podStartSLOduration=4.036691441 podStartE2EDuration="13.670771847s" podCreationTimestamp="2025-12-03 10:10:11 +0000 UTC" firstStartedPulling="2025-12-03 10:10:14.508784739 +0000 UTC m=+5535.077163998" lastFinishedPulling="2025-12-03 10:10:24.142865135 +0000 UTC m=+5544.711244404" observedRunningTime="2025-12-03 10:10:24.665909064 +0000 UTC m=+5545.234288323" watchObservedRunningTime="2025-12-03 10:10:24.670771847 +0000 UTC m=+5545.239151106" Dec 03 10:10:26 crc kubenswrapper[4573]: I1203 10:10:26.031240 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:10:26 crc kubenswrapper[4573]: E1203 10:10:26.031587 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:10:32 crc kubenswrapper[4573]: I1203 10:10:32.255168 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:32 crc kubenswrapper[4573]: I1203 10:10:32.255789 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:32 crc kubenswrapper[4573]: I1203 10:10:32.324155 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:32 crc kubenswrapper[4573]: I1203 10:10:32.795982 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:32 crc kubenswrapper[4573]: I1203 10:10:32.860565 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jklvl"] Dec 03 10:10:34 crc kubenswrapper[4573]: I1203 10:10:34.133910 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-9wr28_7006af64-f3a2-4410-84e4-f6dfd3bcf0ae/nmstate-console-plugin/0.log" Dec 03 10:10:34 crc kubenswrapper[4573]: I1203 10:10:34.357792 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6mwnn_72904583-c577-4d6a-91e1-3ec1d51ec7c2/nmstate-handler/0.log" Dec 03 10:10:34 crc kubenswrapper[4573]: I1203 10:10:34.594003 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-g2mdj_483c74ea-46d0-42e4-84c3-35ed45a0507e/kube-rbac-proxy/0.log" Dec 03 10:10:34 crc kubenswrapper[4573]: I1203 10:10:34.605536 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-g2mdj_483c74ea-46d0-42e4-84c3-35ed45a0507e/nmstate-metrics/0.log" Dec 03 10:10:34 crc kubenswrapper[4573]: I1203 10:10:34.692225 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-j7vz4_d8f81a47-9022-45cf-93e5-8989ee9121ce/nmstate-operator/0.log" Dec 03 10:10:34 crc kubenswrapper[4573]: I1203 10:10:34.762262 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jklvl" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerName="registry-server" containerID="cri-o://8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455" gracePeriod=2 Dec 03 10:10:34 crc kubenswrapper[4573]: I1203 10:10:34.961259 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-hlbnj_358ba50e-2a33-45a2-b1d4-b606638246d5/nmstate-webhook/0.log" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.289659 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.331871 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-utilities\") pod \"16882f5a-571a-471e-92e0-7ff4307fa6c5\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.331970 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-catalog-content\") pod \"16882f5a-571a-471e-92e0-7ff4307fa6c5\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.332103 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h99cp\" (UniqueName: \"kubernetes.io/projected/16882f5a-571a-471e-92e0-7ff4307fa6c5-kube-api-access-h99cp\") pod \"16882f5a-571a-471e-92e0-7ff4307fa6c5\" (UID: \"16882f5a-571a-471e-92e0-7ff4307fa6c5\") " Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.355940 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-utilities" (OuterVolumeSpecName: "utilities") pod "16882f5a-571a-471e-92e0-7ff4307fa6c5" (UID: "16882f5a-571a-471e-92e0-7ff4307fa6c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.356706 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16882f5a-571a-471e-92e0-7ff4307fa6c5-kube-api-access-h99cp" (OuterVolumeSpecName: "kube-api-access-h99cp") pod "16882f5a-571a-471e-92e0-7ff4307fa6c5" (UID: "16882f5a-571a-471e-92e0-7ff4307fa6c5"). InnerVolumeSpecName "kube-api-access-h99cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.434777 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.434820 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h99cp\" (UniqueName: \"kubernetes.io/projected/16882f5a-571a-471e-92e0-7ff4307fa6c5-kube-api-access-h99cp\") on node \"crc\" DevicePath \"\"" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.484149 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16882f5a-571a-471e-92e0-7ff4307fa6c5" (UID: "16882f5a-571a-471e-92e0-7ff4307fa6c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.537349 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16882f5a-571a-471e-92e0-7ff4307fa6c5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.774775 4573 generic.go:334] "Generic (PLEG): container finished" podID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerID="8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455" exitCode=0 Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.774835 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jklvl" event={"ID":"16882f5a-571a-471e-92e0-7ff4307fa6c5","Type":"ContainerDied","Data":"8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455"} Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.774876 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jklvl" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.774910 4573 scope.go:117] "RemoveContainer" containerID="8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.774894 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jklvl" event={"ID":"16882f5a-571a-471e-92e0-7ff4307fa6c5","Type":"ContainerDied","Data":"7d12851de5eb1fb67110b03025c97df52eda495c9f42f246fb4eccac1600f964"} Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.813513 4573 scope.go:117] "RemoveContainer" containerID="40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.887687 4573 scope.go:117] "RemoveContainer" containerID="1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.898462 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jklvl"] Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.925162 4573 scope.go:117] "RemoveContainer" containerID="8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.925523 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jklvl"] Dec 03 10:10:35 crc kubenswrapper[4573]: E1203 10:10:35.938226 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455\": container with ID starting with 8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455 not found: ID does not exist" containerID="8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.938285 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455"} err="failed to get container status \"8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455\": rpc error: code = NotFound desc = could not find container \"8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455\": container with ID starting with 8f83597818173fc73e97f5ff7b518b80c1d24f9e81a288cc8f237f9d83bb8455 not found: ID does not exist" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.938314 4573 scope.go:117] "RemoveContainer" containerID="40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc" Dec 03 10:10:35 crc kubenswrapper[4573]: E1203 10:10:35.941498 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc\": container with ID starting with 40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc not found: ID does not exist" containerID="40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.941567 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc"} err="failed to get container status \"40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc\": rpc error: code = NotFound desc = could not find container \"40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc\": container with ID starting with 40d2be154cefaed44136a133e47be3614e8d2a38ac1e19721e394734c8d448cc not found: ID does not exist" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.941603 4573 scope.go:117] "RemoveContainer" containerID="1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991" Dec 03 10:10:35 crc kubenswrapper[4573]: E1203 10:10:35.944562 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991\": container with ID starting with 1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991 not found: ID does not exist" containerID="1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991" Dec 03 10:10:35 crc kubenswrapper[4573]: I1203 10:10:35.944611 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991"} err="failed to get container status \"1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991\": rpc error: code = NotFound desc = could not find container \"1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991\": container with ID starting with 1a7e6f164876426d64c6447035632219b6c06f456dc8d0416019a331256e5991 not found: ID does not exist" Dec 03 10:10:36 crc kubenswrapper[4573]: I1203 10:10:36.042499 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" path="/var/lib/kubelet/pods/16882f5a-571a-471e-92e0-7ff4307fa6c5/volumes" Dec 03 10:10:41 crc kubenswrapper[4573]: I1203 10:10:41.031171 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:10:41 crc kubenswrapper[4573]: E1203 10:10:41.031825 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:10:51 crc kubenswrapper[4573]: I1203 10:10:51.501912 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rj7p6_974ec28d-5851-4dd3-b992-9920ed8d32de/kube-rbac-proxy/0.log" Dec 03 10:10:51 crc kubenswrapper[4573]: I1203 10:10:51.540058 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-rj7p6_974ec28d-5851-4dd3-b992-9920ed8d32de/controller/0.log" Dec 03 10:10:51 crc kubenswrapper[4573]: I1203 10:10:51.737700 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-frr-files/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.018296 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-reloader/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.144071 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-frr-files/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.170761 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-metrics/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.187719 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-reloader/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.411684 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-frr-files/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.488873 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-reloader/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.553634 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-metrics/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.574929 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-metrics/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.756700 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-frr-files/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.759964 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-reloader/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.881502 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/cp-metrics/0.log" Dec 03 10:10:52 crc kubenswrapper[4573]: I1203 10:10:52.960277 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/controller/0.log" Dec 03 10:10:53 crc kubenswrapper[4573]: I1203 10:10:53.012602 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/frr-metrics/0.log" Dec 03 10:10:53 crc kubenswrapper[4573]: I1203 10:10:53.115740 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/kube-rbac-proxy/0.log" Dec 03 10:10:53 crc kubenswrapper[4573]: I1203 10:10:53.322412 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/kube-rbac-proxy-frr/0.log" Dec 03 10:10:53 crc kubenswrapper[4573]: I1203 10:10:53.407756 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/reloader/0.log" Dec 03 10:10:53 crc kubenswrapper[4573]: I1203 10:10:53.603386 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-cngn6_0e088580-c78b-42cc-b013-bc58bcffa838/frr-k8s-webhook-server/0.log" Dec 03 10:10:53 crc kubenswrapper[4573]: I1203 10:10:53.770162 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-d75bb676c-wwfz6_29c77033-f8ba-44b5-a275-0861c60318bf/manager/0.log" Dec 03 10:10:54 crc kubenswrapper[4573]: I1203 10:10:54.062005 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-b9cfd44fd-mswlh_4e051555-a35f-45f0-ac65-237da26aa296/webhook-server/0.log" Dec 03 10:10:54 crc kubenswrapper[4573]: I1203 10:10:54.355828 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hw9mt_09e37261-797a-4fb1-9ae2-5de662b865e8/kube-rbac-proxy/0.log" Dec 03 10:10:54 crc kubenswrapper[4573]: I1203 10:10:54.982869 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6fjjm_2d98f6ef-9108-408a-8ae6-b1116c434d68/frr/0.log" Dec 03 10:10:55 crc kubenswrapper[4573]: I1203 10:10:55.030175 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:10:55 crc kubenswrapper[4573]: E1203 10:10:55.030444 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:10:55 crc kubenswrapper[4573]: I1203 10:10:55.043697 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hw9mt_09e37261-797a-4fb1-9ae2-5de662b865e8/speaker/0.log" Dec 03 10:11:09 crc kubenswrapper[4573]: I1203 10:11:09.039746 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:11:10 crc kubenswrapper[4573]: I1203 10:11:10.103830 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"530bd17a44c68dd784fe2cf18a8699a09850c79ee45d596b3446920c6348802d"} Dec 03 10:11:10 crc kubenswrapper[4573]: I1203 10:11:10.230776 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/util/0.log" Dec 03 10:11:10 crc kubenswrapper[4573]: I1203 10:11:10.812906 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/util/0.log" Dec 03 10:11:10 crc kubenswrapper[4573]: I1203 10:11:10.869222 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/pull/0.log" Dec 03 10:11:10 crc kubenswrapper[4573]: I1203 10:11:10.904491 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/pull/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.159826 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/extract/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.183509 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/util/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.230948 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fw9869_3e1c6c46-dab7-403c-afd3-3ff5469a2d36/pull/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.405937 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/util/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.557293 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/pull/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.579208 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/util/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.641662 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/pull/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.874854 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/pull/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.875409 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/util/0.log" Dec 03 10:11:11 crc kubenswrapper[4573]: I1203 10:11:11.913361 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83pw6zs_2860a6bf-c83f-4c6d-8d5a-062578bc40aa/extract/0.log" Dec 03 10:11:12 crc kubenswrapper[4573]: I1203 10:11:12.090893 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-utilities/0.log" Dec 03 10:11:12 crc kubenswrapper[4573]: I1203 10:11:12.332554 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-content/0.log" Dec 03 10:11:12 crc kubenswrapper[4573]: I1203 10:11:12.352739 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-utilities/0.log" Dec 03 10:11:12 crc kubenswrapper[4573]: I1203 10:11:12.386810 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-content/0.log" Dec 03 10:11:12 crc kubenswrapper[4573]: I1203 10:11:12.594316 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-utilities/0.log" Dec 03 10:11:12 crc kubenswrapper[4573]: I1203 10:11:12.646423 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/extract-content/0.log" Dec 03 10:11:12 crc kubenswrapper[4573]: I1203 10:11:12.982997 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-utilities/0.log" Dec 03 10:11:12 crc kubenswrapper[4573]: I1203 10:11:12.994441 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jttp2_cdd548d8-d599-4da1-b815-287a6a4abdab/registry-server/0.log" Dec 03 10:11:13 crc kubenswrapper[4573]: I1203 10:11:13.195031 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-utilities/0.log" Dec 03 10:11:13 crc kubenswrapper[4573]: I1203 10:11:13.239067 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-content/0.log" Dec 03 10:11:13 crc kubenswrapper[4573]: I1203 10:11:13.285717 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-content/0.log" Dec 03 10:11:13 crc kubenswrapper[4573]: I1203 10:11:13.474401 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-content/0.log" Dec 03 10:11:13 crc kubenswrapper[4573]: I1203 10:11:13.498304 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/extract-utilities/0.log" Dec 03 10:11:13 crc kubenswrapper[4573]: I1203 10:11:13.782031 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hgwmj_e82d702a-82de-47e9-89fe-8ca23ce8a98b/marketplace-operator/0.log" Dec 03 10:11:13 crc kubenswrapper[4573]: I1203 10:11:13.918106 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-utilities/0.log" Dec 03 10:11:14 crc kubenswrapper[4573]: I1203 10:11:14.245620 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-content/0.log" Dec 03 10:11:14 crc kubenswrapper[4573]: I1203 10:11:14.299552 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5rtb9_577aa218-59b2-4cbc-ae72-59c607006fb5/registry-server/0.log" Dec 03 10:11:14 crc kubenswrapper[4573]: I1203 10:11:14.337718 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-content/0.log" Dec 03 10:11:14 crc kubenswrapper[4573]: I1203 10:11:14.344381 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-utilities/0.log" Dec 03 10:11:14 crc kubenswrapper[4573]: I1203 10:11:14.607387 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-content/0.log" Dec 03 10:11:14 crc kubenswrapper[4573]: I1203 10:11:14.635917 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/extract-utilities/0.log" Dec 03 10:11:14 crc kubenswrapper[4573]: I1203 10:11:14.801872 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-6v4g4_caa010bb-5d5b-45f6-98b6-d2431c39a5fe/registry-server/0.log" Dec 03 10:11:14 crc kubenswrapper[4573]: I1203 10:11:14.916298 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-utilities/0.log" Dec 03 10:11:15 crc kubenswrapper[4573]: I1203 10:11:15.124866 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-content/0.log" Dec 03 10:11:15 crc kubenswrapper[4573]: I1203 10:11:15.158690 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-utilities/0.log" Dec 03 10:11:15 crc kubenswrapper[4573]: I1203 10:11:15.235742 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-content/0.log" Dec 03 10:11:15 crc kubenswrapper[4573]: I1203 10:11:15.460910 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-utilities/0.log" Dec 03 10:11:15 crc kubenswrapper[4573]: I1203 10:11:15.477963 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/extract-content/0.log" Dec 03 10:11:16 crc kubenswrapper[4573]: I1203 10:11:16.245358 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-tmzcx_f4d1c7fd-3939-4391-ae35-24bc7c2beaeb/registry-server/0.log" Dec 03 10:13:26 crc kubenswrapper[4573]: I1203 10:13:26.943556 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:13:26 crc kubenswrapper[4573]: I1203 10:13:26.944158 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:13:44 crc kubenswrapper[4573]: I1203 10:13:44.802090 4573 generic.go:334] "Generic (PLEG): container finished" podID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerID="0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18" exitCode=0 Dec 03 10:13:44 crc kubenswrapper[4573]: I1203 10:13:44.802532 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t6nkm/must-gather-wfn97" event={"ID":"2521d5a5-1671-42d0-bdd1-180b57a9287b","Type":"ContainerDied","Data":"0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18"} Dec 03 10:13:44 crc kubenswrapper[4573]: I1203 10:13:44.805694 4573 scope.go:117] "RemoveContainer" containerID="0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18" Dec 03 10:13:45 crc kubenswrapper[4573]: I1203 10:13:45.317455 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6nkm_must-gather-wfn97_2521d5a5-1671-42d0-bdd1-180b57a9287b/gather/0.log" Dec 03 10:13:56 crc kubenswrapper[4573]: I1203 10:13:56.943779 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:13:56 crc kubenswrapper[4573]: I1203 10:13:56.944549 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:13:59 crc kubenswrapper[4573]: I1203 10:13:59.929176 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t6nkm/must-gather-wfn97"] Dec 03 10:13:59 crc kubenswrapper[4573]: I1203 10:13:59.929726 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-t6nkm/must-gather-wfn97" podUID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerName="copy" containerID="cri-o://5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700" gracePeriod=2 Dec 03 10:13:59 crc kubenswrapper[4573]: I1203 10:13:59.940701 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t6nkm/must-gather-wfn97"] Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.650866 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6nkm_must-gather-wfn97_2521d5a5-1671-42d0-bdd1-180b57a9287b/copy/0.log" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.651788 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.790897 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2521d5a5-1671-42d0-bdd1-180b57a9287b-must-gather-output\") pod \"2521d5a5-1671-42d0-bdd1-180b57a9287b\" (UID: \"2521d5a5-1671-42d0-bdd1-180b57a9287b\") " Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.791174 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ntmp\" (UniqueName: \"kubernetes.io/projected/2521d5a5-1671-42d0-bdd1-180b57a9287b-kube-api-access-7ntmp\") pod \"2521d5a5-1671-42d0-bdd1-180b57a9287b\" (UID: \"2521d5a5-1671-42d0-bdd1-180b57a9287b\") " Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.798164 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2521d5a5-1671-42d0-bdd1-180b57a9287b-kube-api-access-7ntmp" (OuterVolumeSpecName: "kube-api-access-7ntmp") pod "2521d5a5-1671-42d0-bdd1-180b57a9287b" (UID: "2521d5a5-1671-42d0-bdd1-180b57a9287b"). InnerVolumeSpecName "kube-api-access-7ntmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.894890 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ntmp\" (UniqueName: \"kubernetes.io/projected/2521d5a5-1671-42d0-bdd1-180b57a9287b-kube-api-access-7ntmp\") on node \"crc\" DevicePath \"\"" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.972116 4573 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t6nkm_must-gather-wfn97_2521d5a5-1671-42d0-bdd1-180b57a9287b/copy/0.log" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.973601 4573 generic.go:334] "Generic (PLEG): container finished" podID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerID="5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700" exitCode=143 Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.973649 4573 scope.go:117] "RemoveContainer" containerID="5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.973700 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t6nkm/must-gather-wfn97" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.987189 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2521d5a5-1671-42d0-bdd1-180b57a9287b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2521d5a5-1671-42d0-bdd1-180b57a9287b" (UID: "2521d5a5-1671-42d0-bdd1-180b57a9287b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.997393 4573 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2521d5a5-1671-42d0-bdd1-180b57a9287b-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 10:14:00 crc kubenswrapper[4573]: I1203 10:14:00.997714 4573 scope.go:117] "RemoveContainer" containerID="0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18" Dec 03 10:14:01 crc kubenswrapper[4573]: I1203 10:14:01.044913 4573 scope.go:117] "RemoveContainer" containerID="5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700" Dec 03 10:14:01 crc kubenswrapper[4573]: E1203 10:14:01.045485 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700\": container with ID starting with 5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700 not found: ID does not exist" containerID="5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700" Dec 03 10:14:01 crc kubenswrapper[4573]: I1203 10:14:01.045559 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700"} err="failed to get container status \"5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700\": rpc error: code = NotFound desc = could not find container \"5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700\": container with ID starting with 5a743f85d56b0da32fbac0a26b9d894fd487e9bd8bac0e9ca9a8a7f149c31700 not found: ID does not exist" Dec 03 10:14:01 crc kubenswrapper[4573]: I1203 10:14:01.045589 4573 scope.go:117] "RemoveContainer" containerID="0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18" Dec 03 10:14:01 crc kubenswrapper[4573]: E1203 10:14:01.045997 4573 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18\": container with ID starting with 0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18 not found: ID does not exist" containerID="0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18" Dec 03 10:14:01 crc kubenswrapper[4573]: I1203 10:14:01.046024 4573 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18"} err="failed to get container status \"0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18\": rpc error: code = NotFound desc = could not find container \"0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18\": container with ID starting with 0fd6fe761aa0c7d03f68f1b789d9365299b89790f5878b5bc2a751860a3b5e18 not found: ID does not exist" Dec 03 10:14:02 crc kubenswrapper[4573]: I1203 10:14:02.139703 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2521d5a5-1671-42d0-bdd1-180b57a9287b" path="/var/lib/kubelet/pods/2521d5a5-1671-42d0-bdd1-180b57a9287b/volumes" Dec 03 10:14:10 crc kubenswrapper[4573]: I1203 10:14:10.284364 4573 scope.go:117] "RemoveContainer" containerID="b418b1c5d229cffa0dcb78723a82c7821ed23e422d0cb8a5782d6c225a3e06a7" Dec 03 10:14:26 crc kubenswrapper[4573]: I1203 10:14:26.943262 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:14:26 crc kubenswrapper[4573]: I1203 10:14:26.943697 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:14:26 crc kubenswrapper[4573]: I1203 10:14:26.943758 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 10:14:26 crc kubenswrapper[4573]: I1203 10:14:26.945522 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"530bd17a44c68dd784fe2cf18a8699a09850c79ee45d596b3446920c6348802d"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 10:14:26 crc kubenswrapper[4573]: I1203 10:14:26.945597 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://530bd17a44c68dd784fe2cf18a8699a09850c79ee45d596b3446920c6348802d" gracePeriod=600 Dec 03 10:14:27 crc kubenswrapper[4573]: I1203 10:14:27.250472 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="530bd17a44c68dd784fe2cf18a8699a09850c79ee45d596b3446920c6348802d" exitCode=0 Dec 03 10:14:27 crc kubenswrapper[4573]: I1203 10:14:27.250517 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"530bd17a44c68dd784fe2cf18a8699a09850c79ee45d596b3446920c6348802d"} Dec 03 10:14:27 crc kubenswrapper[4573]: I1203 10:14:27.250554 4573 scope.go:117] "RemoveContainer" containerID="87c512bfcaa736348310c432643d569c93206c08c6ecf6d6ab56ebd39025d9d2" Dec 03 10:14:28 crc kubenswrapper[4573]: I1203 10:14:28.264143 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerStarted","Data":"299978d43b45f39151155b9e9635a322c63747d8ccc0c6bfe04ed2022d8f904d"} Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.156829 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2"] Dec 03 10:15:00 crc kubenswrapper[4573]: E1203 10:15:00.158018 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerName="copy" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.158038 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerName="copy" Dec 03 10:15:00 crc kubenswrapper[4573]: E1203 10:15:00.158226 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerName="gather" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.158238 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerName="gather" Dec 03 10:15:00 crc kubenswrapper[4573]: E1203 10:15:00.158266 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerName="extract-content" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.158275 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerName="extract-content" Dec 03 10:15:00 crc kubenswrapper[4573]: E1203 10:15:00.158299 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerName="extract-utilities" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.158307 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerName="extract-utilities" Dec 03 10:15:00 crc kubenswrapper[4573]: E1203 10:15:00.158330 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerName="registry-server" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.158338 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerName="registry-server" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.158630 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerName="gather" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.158670 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="16882f5a-571a-471e-92e0-7ff4307fa6c5" containerName="registry-server" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.158685 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="2521d5a5-1671-42d0-bdd1-180b57a9287b" containerName="copy" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.159666 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.163268 4573 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.164139 4573 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.168165 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2"] Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.287735 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56vdr\" (UniqueName: \"kubernetes.io/projected/cd57dcac-c6df-40d2-97a5-120bc12b5d69-kube-api-access-56vdr\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.287833 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd57dcac-c6df-40d2-97a5-120bc12b5d69-config-volume\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.287908 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd57dcac-c6df-40d2-97a5-120bc12b5d69-secret-volume\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.390083 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd57dcac-c6df-40d2-97a5-120bc12b5d69-secret-volume\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.390263 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56vdr\" (UniqueName: \"kubernetes.io/projected/cd57dcac-c6df-40d2-97a5-120bc12b5d69-kube-api-access-56vdr\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.390315 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd57dcac-c6df-40d2-97a5-120bc12b5d69-config-volume\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.393961 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd57dcac-c6df-40d2-97a5-120bc12b5d69-config-volume\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.405123 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd57dcac-c6df-40d2-97a5-120bc12b5d69-secret-volume\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.417788 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56vdr\" (UniqueName: \"kubernetes.io/projected/cd57dcac-c6df-40d2-97a5-120bc12b5d69-kube-api-access-56vdr\") pod \"collect-profiles-29412615-p4gk2\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.491894 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:00 crc kubenswrapper[4573]: I1203 10:15:00.969347 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2"] Dec 03 10:15:00 crc kubenswrapper[4573]: W1203 10:15:00.972735 4573 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd57dcac_c6df_40d2_97a5_120bc12b5d69.slice/crio-9b07ed079288b7ae7f6efb0cbe54b7dd896cf248bd40d36b2808b0680e633520 WatchSource:0}: Error finding container 9b07ed079288b7ae7f6efb0cbe54b7dd896cf248bd40d36b2808b0680e633520: Status 404 returned error can't find the container with id 9b07ed079288b7ae7f6efb0cbe54b7dd896cf248bd40d36b2808b0680e633520 Dec 03 10:15:01 crc kubenswrapper[4573]: I1203 10:15:01.607220 4573 generic.go:334] "Generic (PLEG): container finished" podID="cd57dcac-c6df-40d2-97a5-120bc12b5d69" containerID="ba5c66185fcb635a8dd622aff4e015be0e2a9577bce935fd9c76f3c97cc414ec" exitCode=0 Dec 03 10:15:01 crc kubenswrapper[4573]: I1203 10:15:01.607334 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" event={"ID":"cd57dcac-c6df-40d2-97a5-120bc12b5d69","Type":"ContainerDied","Data":"ba5c66185fcb635a8dd622aff4e015be0e2a9577bce935fd9c76f3c97cc414ec"} Dec 03 10:15:01 crc kubenswrapper[4573]: I1203 10:15:01.607553 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" event={"ID":"cd57dcac-c6df-40d2-97a5-120bc12b5d69","Type":"ContainerStarted","Data":"9b07ed079288b7ae7f6efb0cbe54b7dd896cf248bd40d36b2808b0680e633520"} Dec 03 10:15:02 crc kubenswrapper[4573]: I1203 10:15:02.987035 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.158621 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd57dcac-c6df-40d2-97a5-120bc12b5d69-config-volume\") pod \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.158949 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd57dcac-c6df-40d2-97a5-120bc12b5d69-secret-volume\") pod \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.159094 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56vdr\" (UniqueName: \"kubernetes.io/projected/cd57dcac-c6df-40d2-97a5-120bc12b5d69-kube-api-access-56vdr\") pod \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\" (UID: \"cd57dcac-c6df-40d2-97a5-120bc12b5d69\") " Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.160603 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd57dcac-c6df-40d2-97a5-120bc12b5d69-config-volume" (OuterVolumeSpecName: "config-volume") pod "cd57dcac-c6df-40d2-97a5-120bc12b5d69" (UID: "cd57dcac-c6df-40d2-97a5-120bc12b5d69"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.166187 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd57dcac-c6df-40d2-97a5-120bc12b5d69-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cd57dcac-c6df-40d2-97a5-120bc12b5d69" (UID: "cd57dcac-c6df-40d2-97a5-120bc12b5d69"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.167250 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd57dcac-c6df-40d2-97a5-120bc12b5d69-kube-api-access-56vdr" (OuterVolumeSpecName: "kube-api-access-56vdr") pod "cd57dcac-c6df-40d2-97a5-120bc12b5d69" (UID: "cd57dcac-c6df-40d2-97a5-120bc12b5d69"). InnerVolumeSpecName "kube-api-access-56vdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.261553 4573 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd57dcac-c6df-40d2-97a5-120bc12b5d69-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.261598 4573 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd57dcac-c6df-40d2-97a5-120bc12b5d69-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.261610 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56vdr\" (UniqueName: \"kubernetes.io/projected/cd57dcac-c6df-40d2-97a5-120bc12b5d69-kube-api-access-56vdr\") on node \"crc\" DevicePath \"\"" Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.629452 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" event={"ID":"cd57dcac-c6df-40d2-97a5-120bc12b5d69","Type":"ContainerDied","Data":"9b07ed079288b7ae7f6efb0cbe54b7dd896cf248bd40d36b2808b0680e633520"} Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.629500 4573 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b07ed079288b7ae7f6efb0cbe54b7dd896cf248bd40d36b2808b0680e633520" Dec 03 10:15:03 crc kubenswrapper[4573]: I1203 10:15:03.629571 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412615-p4gk2" Dec 03 10:15:04 crc kubenswrapper[4573]: I1203 10:15:04.083638 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn"] Dec 03 10:15:04 crc kubenswrapper[4573]: I1203 10:15:04.093933 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412570-q7ssn"] Dec 03 10:15:06 crc kubenswrapper[4573]: I1203 10:15:06.044569 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46e2a8e9-4614-4763-979a-daa41262f708" path="/var/lib/kubelet/pods/46e2a8e9-4614-4763-979a-daa41262f708/volumes" Dec 03 10:15:10 crc kubenswrapper[4573]: I1203 10:15:10.356637 4573 scope.go:117] "RemoveContainer" containerID="29212cc6ebedf6eb1bdc4823a7236850851d7da785826f168634e249af29ed1b" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.751619 4573 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5sszq"] Dec 03 10:16:20 crc kubenswrapper[4573]: E1203 10:16:20.752633 4573 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd57dcac-c6df-40d2-97a5-120bc12b5d69" containerName="collect-profiles" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.752646 4573 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd57dcac-c6df-40d2-97a5-120bc12b5d69" containerName="collect-profiles" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.752859 4573 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd57dcac-c6df-40d2-97a5-120bc12b5d69" containerName="collect-profiles" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.754287 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.763559 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5sszq"] Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.886408 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t95fx\" (UniqueName: \"kubernetes.io/projected/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-kube-api-access-t95fx\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.886617 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-catalog-content\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.886777 4573 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-utilities\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.988834 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t95fx\" (UniqueName: \"kubernetes.io/projected/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-kube-api-access-t95fx\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.988962 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-catalog-content\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.989031 4573 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-utilities\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.989602 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-catalog-content\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:20 crc kubenswrapper[4573]: I1203 10:16:20.989787 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-utilities\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:21 crc kubenswrapper[4573]: I1203 10:16:21.015272 4573 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t95fx\" (UniqueName: \"kubernetes.io/projected/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-kube-api-access-t95fx\") pod \"community-operators-5sszq\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:21 crc kubenswrapper[4573]: I1203 10:16:21.078909 4573 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:21 crc kubenswrapper[4573]: I1203 10:16:21.754085 4573 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5sszq"] Dec 03 10:16:22 crc kubenswrapper[4573]: I1203 10:16:22.656902 4573 generic.go:334] "Generic (PLEG): container finished" podID="a40d7247-bdd3-4fbb-8f88-3a62887bd9c7" containerID="aab5955fecec2afef81afc5d83c183d82354e49e197f6b08cc27a253e50f565c" exitCode=0 Dec 03 10:16:22 crc kubenswrapper[4573]: I1203 10:16:22.656964 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5sszq" event={"ID":"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7","Type":"ContainerDied","Data":"aab5955fecec2afef81afc5d83c183d82354e49e197f6b08cc27a253e50f565c"} Dec 03 10:16:22 crc kubenswrapper[4573]: I1203 10:16:22.657317 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5sszq" event={"ID":"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7","Type":"ContainerStarted","Data":"99ce3272e71bf84eeebf5d88d2df2d2d786c564156014be96991f05526b3c7f7"} Dec 03 10:16:22 crc kubenswrapper[4573]: I1203 10:16:22.660753 4573 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 10:16:23 crc kubenswrapper[4573]: I1203 10:16:23.670351 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5sszq" event={"ID":"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7","Type":"ContainerStarted","Data":"2346982df3bbf01e5a979700c905bd8e1c5d81658cecc9a29bd709ba869e3e87"} Dec 03 10:16:24 crc kubenswrapper[4573]: I1203 10:16:24.684749 4573 generic.go:334] "Generic (PLEG): container finished" podID="a40d7247-bdd3-4fbb-8f88-3a62887bd9c7" containerID="2346982df3bbf01e5a979700c905bd8e1c5d81658cecc9a29bd709ba869e3e87" exitCode=0 Dec 03 10:16:24 crc kubenswrapper[4573]: I1203 10:16:24.684799 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5sszq" event={"ID":"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7","Type":"ContainerDied","Data":"2346982df3bbf01e5a979700c905bd8e1c5d81658cecc9a29bd709ba869e3e87"} Dec 03 10:16:25 crc kubenswrapper[4573]: I1203 10:16:25.696306 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5sszq" event={"ID":"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7","Type":"ContainerStarted","Data":"e47c517a4509aa82f16d9dc3961195dd5713e4870e2ee94732b741c3b4a59038"} Dec 03 10:16:25 crc kubenswrapper[4573]: I1203 10:16:25.730302 4573 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5sszq" podStartSLOduration=3.218732825 podStartE2EDuration="5.73028432s" podCreationTimestamp="2025-12-03 10:16:20 +0000 UTC" firstStartedPulling="2025-12-03 10:16:22.660355924 +0000 UTC m=+5903.228735193" lastFinishedPulling="2025-12-03 10:16:25.171907429 +0000 UTC m=+5905.740286688" observedRunningTime="2025-12-03 10:16:25.725611452 +0000 UTC m=+5906.293990711" watchObservedRunningTime="2025-12-03 10:16:25.73028432 +0000 UTC m=+5906.298663579" Dec 03 10:16:31 crc kubenswrapper[4573]: I1203 10:16:31.080108 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:31 crc kubenswrapper[4573]: I1203 10:16:31.080673 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:31 crc kubenswrapper[4573]: I1203 10:16:31.131085 4573 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:31 crc kubenswrapper[4573]: I1203 10:16:31.839635 4573 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:31 crc kubenswrapper[4573]: I1203 10:16:31.892939 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5sszq"] Dec 03 10:16:33 crc kubenswrapper[4573]: I1203 10:16:33.794571 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5sszq" podUID="a40d7247-bdd3-4fbb-8f88-3a62887bd9c7" containerName="registry-server" containerID="cri-o://e47c517a4509aa82f16d9dc3961195dd5713e4870e2ee94732b741c3b4a59038" gracePeriod=2 Dec 03 10:16:34 crc kubenswrapper[4573]: I1203 10:16:34.805383 4573 generic.go:334] "Generic (PLEG): container finished" podID="a40d7247-bdd3-4fbb-8f88-3a62887bd9c7" containerID="e47c517a4509aa82f16d9dc3961195dd5713e4870e2ee94732b741c3b4a59038" exitCode=0 Dec 03 10:16:34 crc kubenswrapper[4573]: I1203 10:16:34.805456 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5sszq" event={"ID":"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7","Type":"ContainerDied","Data":"e47c517a4509aa82f16d9dc3961195dd5713e4870e2ee94732b741c3b4a59038"} Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.101621 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.214631 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-catalog-content\") pod \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.214704 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t95fx\" (UniqueName: \"kubernetes.io/projected/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-kube-api-access-t95fx\") pod \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.214851 4573 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-utilities\") pod \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\" (UID: \"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7\") " Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.215627 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-utilities" (OuterVolumeSpecName: "utilities") pod "a40d7247-bdd3-4fbb-8f88-3a62887bd9c7" (UID: "a40d7247-bdd3-4fbb-8f88-3a62887bd9c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.223127 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-kube-api-access-t95fx" (OuterVolumeSpecName: "kube-api-access-t95fx") pod "a40d7247-bdd3-4fbb-8f88-3a62887bd9c7" (UID: "a40d7247-bdd3-4fbb-8f88-3a62887bd9c7"). InnerVolumeSpecName "kube-api-access-t95fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.297506 4573 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a40d7247-bdd3-4fbb-8f88-3a62887bd9c7" (UID: "a40d7247-bdd3-4fbb-8f88-3a62887bd9c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.316966 4573 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t95fx\" (UniqueName: \"kubernetes.io/projected/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-kube-api-access-t95fx\") on node \"crc\" DevicePath \"\"" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.317006 4573 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.317016 4573 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.816877 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5sszq" event={"ID":"a40d7247-bdd3-4fbb-8f88-3a62887bd9c7","Type":"ContainerDied","Data":"99ce3272e71bf84eeebf5d88d2df2d2d786c564156014be96991f05526b3c7f7"} Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.816929 4573 scope.go:117] "RemoveContainer" containerID="e47c517a4509aa82f16d9dc3961195dd5713e4870e2ee94732b741c3b4a59038" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.816989 4573 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5sszq" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.847275 4573 scope.go:117] "RemoveContainer" containerID="2346982df3bbf01e5a979700c905bd8e1c5d81658cecc9a29bd709ba869e3e87" Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.873192 4573 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5sszq"] Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.896948 4573 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5sszq"] Dec 03 10:16:35 crc kubenswrapper[4573]: I1203 10:16:35.902560 4573 scope.go:117] "RemoveContainer" containerID="aab5955fecec2afef81afc5d83c183d82354e49e197f6b08cc27a253e50f565c" Dec 03 10:16:36 crc kubenswrapper[4573]: I1203 10:16:36.046240 4573 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a40d7247-bdd3-4fbb-8f88-3a62887bd9c7" path="/var/lib/kubelet/pods/a40d7247-bdd3-4fbb-8f88-3a62887bd9c7/volumes" Dec 03 10:16:56 crc kubenswrapper[4573]: I1203 10:16:56.943143 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:16:56 crc kubenswrapper[4573]: I1203 10:16:56.943718 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:17:26 crc kubenswrapper[4573]: I1203 10:17:26.943408 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:17:26 crc kubenswrapper[4573]: I1203 10:17:26.943858 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:17:56 crc kubenswrapper[4573]: I1203 10:17:56.942993 4573 patch_prober.go:28] interesting pod/machine-config-daemon-xg4ms container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 10:17:56 crc kubenswrapper[4573]: I1203 10:17:56.943465 4573 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 10:17:56 crc kubenswrapper[4573]: I1203 10:17:56.943517 4573 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" Dec 03 10:17:56 crc kubenswrapper[4573]: I1203 10:17:56.944368 4573 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"299978d43b45f39151155b9e9635a322c63747d8ccc0c6bfe04ed2022d8f904d"} pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 10:17:56 crc kubenswrapper[4573]: I1203 10:17:56.944434 4573 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" containerName="machine-config-daemon" containerID="cri-o://299978d43b45f39151155b9e9635a322c63747d8ccc0c6bfe04ed2022d8f904d" gracePeriod=600 Dec 03 10:17:57 crc kubenswrapper[4573]: E1203 10:17:57.090371 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" Dec 03 10:17:57 crc kubenswrapper[4573]: I1203 10:17:57.660677 4573 generic.go:334] "Generic (PLEG): container finished" podID="e17325e8-2a46-4028-b7c5-664eeda96013" containerID="299978d43b45f39151155b9e9635a322c63747d8ccc0c6bfe04ed2022d8f904d" exitCode=0 Dec 03 10:17:57 crc kubenswrapper[4573]: I1203 10:17:57.660954 4573 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" event={"ID":"e17325e8-2a46-4028-b7c5-664eeda96013","Type":"ContainerDied","Data":"299978d43b45f39151155b9e9635a322c63747d8ccc0c6bfe04ed2022d8f904d"} Dec 03 10:17:57 crc kubenswrapper[4573]: I1203 10:17:57.661012 4573 scope.go:117] "RemoveContainer" containerID="530bd17a44c68dd784fe2cf18a8699a09850c79ee45d596b3446920c6348802d" Dec 03 10:17:57 crc kubenswrapper[4573]: I1203 10:17:57.661907 4573 scope.go:117] "RemoveContainer" containerID="299978d43b45f39151155b9e9635a322c63747d8ccc0c6bfe04ed2022d8f904d" Dec 03 10:17:57 crc kubenswrapper[4573]: E1203 10:17:57.662362 4573 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xg4ms_openshift-machine-config-operator(e17325e8-2a46-4028-b7c5-664eeda96013)\"" pod="openshift-machine-config-operator/machine-config-daemon-xg4ms" podUID="e17325e8-2a46-4028-b7c5-664eeda96013" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114007135024442 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114007135017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113772752016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113772752015467 5ustar corecore